Next Article in Journal
A Review of the Literature about Sustainability in the Work of the Future: An Overview of Industry 4.0 and Human Resources
Next Article in Special Issue
Towards a Domain-Neutral Platform for Sustainable Digital Twin Development
Previous Article in Journal
Study of the Spatial Spillover Effects of the Efficiency of Agricultural Product Circulation in Provinces along the Belt and Road under the Green Total Factor Productivity Framework
Previous Article in Special Issue
Smart Agricultural–Industrial Crop-Monitoring System Using Unmanned Aerial Vehicle–Internet of Things Classification Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intrusion Detection in Healthcare 4.0 Internet of Things Systems via Metaheuristics Optimized Machine Learning

1
Faculty of Informatics and Computing, Singidunum University, 11010 Belgrade, Serbia
2
Teacher Education Faculty, University of Pristina in Kosovska Mitrovica, 38220 Kosovska Mitrovica, Serbia
3
Department of Applied Informatics, Vytautas Magnus University, 44404 Kaunas, Lithuania
4
School of Electrical Engineering, University of Belgrade, 11000 Belgrade, Serbia
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(16), 12563; https://doi.org/10.3390/su151612563
Submission received: 24 July 2023 / Revised: 12 August 2023 / Accepted: 15 August 2023 / Published: 18 August 2023
(This article belongs to the Special Issue Sustainable Information Engineering and Computer Science)

Abstract

:
Rapid developments in Internet of Things (IoT) systems have led to a wide integration of such systems into everyday life. Systems for active real-time monitoring are especially useful in areas where rapid action can have a significant impact on outcomes such as healthcare. However, a major challenge persists within IoT that limit wider integration. Sustainable healthcare supported by the IoT must provide organized healthcare to the population, without compromising the environment. Security plays a major role in the sustainability of IoT systems, therefore detecting and taking timely action is one step in overcoming the sustainability challenges. This work tackles security challenges head-on through the use of machine learning algorithms optimized via a modified Firefly algorithm for detecting security issues in IoT devices used for Healthcare 4.0. Metaheuristic solutions have contributed to sustainability in various areas as they can solve nondeterministic polynomial time-hard problem (NP-hard) problems in realistic time and with accuracy which are paramount for sustainable systems in any sector and especially in healthcare. Experiments on a synthetic dataset generated by an advanced configuration tool for IoT structures are performed. Also, multiple well-known machine learning models were used and optimized by introducing modified firefly metaheuristics. The best models have been subjected to SHapley Additive exPlanations (SHAP) analysis to determine the factors that contribute to occurring issues. Conclusions from all the performed testing and comparisons indicate significant improvements in the formulated problem.

1. Introduction

Healthcare 4.0 [1] is essentially a concept that can be considered to be a part of the Fourth Industrial Revolution, and it can be implemented by utilizing technologies such as artificial intelligence (AI), IoT, data analytics, and many others. This concept is crucial for overall improvement in the healthcare industry, enabling better accessibility, enhancing quality, and ensuring high efficiency in healthcare treatment [1]. By using this concept, the integration of personalized medicine is achieved, and smart patient care devices can be implemented, along with remote healthcare delivery solutions which are all paramount to achieving a sustainable healthcare system. Advanced technologies like Healthcare 4.0 can have certain challenges, particularly in terms of data security and privacy [2], which hinder its adoption rate. The primary concern revolves around protecting user data [3], which contains sensitive information about the service user or patient. The data stored in the database is of an extremely sensitive nature, encompassing personal information such as the user’s genetic code, medical history, identification details, and many others. Additionally, the lack of standards within this concept poses integration issues with various systems. Introducing standards can improve the mechanism of data exchange among different device levels and facilitate resource sharing among relevant healthcare institutions.
An important aspect of the Healthcare 4.0 concept which heavily influences sustainability is the provision of real-time information, aiming to enable faster and more efficient responses from medical staff and doctors, thereby achieving more effective patient diagnoses [1]. At any given moment, doctors have access to data indicating the current condition of the patient. It is also noteworthy that vital patient parameters can be monitored live, in real time [4]. In the event of deviations from the defined parameters set by the doctor, the medical staff and the doctor can immediately change the therapy and provide the appropriate treatment. Additionally, the Healthcare 4.0 concept facilitates the efficiency of detecting emergency cases, such as a heart attack in a patient. In such cases, appropriate measures are taken to save the patient’s life. The technology of Healthcare 4.0 can facilitate consultations with doctors in relevant medical fields, allowing for a more efficient exchange of necessary information with relevant medical institutions. This enables faster medical decision-making based on real-time diagnostics. The influence on sustainable cities is present in Healthcare 4.0 as well. Route optimization is important for sustainable health service providing which intertwines with other concepts like smart and sustainable cities. There is more on this topic in Section 2, which provides the literature review.
Every technology in the world has its vulnerabilities, and Healthcare 4.0 [3], with the use of the IoT, also faces issues regarding data protection and integrity. One of the attack techniques present in this technology is the man-in-the-middle attack, where patient data are intercepted. Furthermore, due to the use of weaker data encryption techniques [5] in the exchange of data between patients and relevant institutions on IoT devices, unauthorized access to sensitive and personal patient information can occur. Computer networks and systems that have IoT devices and integrated Healthcare 4.0 technology on their servers can be susceptible to potential attacks on the network infrastructure, where attackers intentionally attempt to disrupt the use of active services. Such attacks can cause damage to the network infrastructure or render it completely incapacitated [6,7]. As a result of such attacks, the attacker can install malicious software on devices, aiming to target IoT devices that interact with patients. The reason for these attacks is unauthorized access to patient data, where even solutions like blockchain are combined with IoT structures [8]. In addition to the proper training of employees regarding potential data leaks, where the human factor often plays a significant role, possible protection against such attacks can be the implementation and integration of hardware and software firewall mechanisms, as well as the implementation of the intrusion detection system (IDS) and intrusion prevention system (IPS) [9]. The implementation of IDS/IPS systems is of great importance to detect any suspicious activity in computer networks or cloud systems where Healthcare 4.0 technology is implemented in a timely manner [9]. Conclusively, a sustainable health system is not possible without ensuring a stable and secure infrastructure which is challenging as is due to a large number of device types and devices in general.
Healthcare 4.0 technology requires the implementation of mechanisms that can adapt to new potential vulnerabilities. In addition to the previously mentioned systems, other technologies can be utilized, such as automated incident response systems, security analytics systems, and identity and access management systems. Furthermore, a vulnerability management system can be employed to identify issues and detect vulnerabilities within the infrastructure of computer networks or cloud systems.
Artificial intelligence holds great potential in addressing numerous vulnerabilities and overall security challenges faced by Healthcare 4.0 technology. AI can utilize advanced machine learning algorithms to analyze large datasets and draw conclusions about potential threats or attacks. From such data sets, it is possible to identify patterns that may indicate a threat or attack. Moreover, based on such analyses, AI can discover and detect new potential system vulnerabilities that may appear unusual to the implemented systems. By leveraging known vulnerabilities, AI can even predict or anticipate attacks in advance.
Advanced machine learning algorithms can predict or take preventive measures before an actual infrastructure attack occurs. Through automation, efficiency can be achieved by implementing automated incident response in the infrastructure [9]. In the event of an attack or the introduction of malicious code into the infrastructure, AI can recognize attack patterns and automatically respond to them [9].
When it comes to data within the infrastructure where user data are stored, AI can enhance encryption mechanisms using deep learning on existing encryption algorithms. This improves the existing data encryption technique, and through deep learning, vulnerability detection mechanisms and systems can be enhanced, resulting in more modern data encryption systems being produced automatically.
To achieve sustainable environments and higher quality of life, it is paramount to take into account all the different aspects of the healthcare systems that are not universal around the globe. Researchers explore topics like the supply chains of pharmaceuticals [10], combine technologies like blockchain with IoT infrastructure for the optimized road health assistance [11], and even topics that do not affect the patient directly but electricity management in a hospital can bring other types of quality of service improvements and increase the well-being of patients leading the society to a more sustainable world [12].
Extreme gradient boosting (XGBoost) is an algorithm that has proven to be the best optimizer for competition. It is a gradient-boosting-based approach that increases the learner’s performance and reduces bias and even variance. It is considered a tree-based algorithm due to its foundation being strongly related to it. Due to the complicated process of hyperparameter optimization, XGBoost is paired with an optimizer.
Metaheuristic optimization is recognized to be on the rise because of the high-performance outputs that can be achieved. The advantage of such solutions comes from their NP hard-solving capabilities. The need for extensive experimentation in this field comes from the no free lunch (NFL) theorem, and according to it, no single solution is equally good for solving all problems [13].
The authors have chosen the firefly algorithm (FA) metaheuristic based on swarm behavior as a method for optimization due to its history of high performance for NP-hard optimization. Even though they are powerful optimizers, swarm intelligence (SI) algorithms are not without shortcomings. Where one algorithm excels, another fails. Hence, the authors apply modifications to the original solution, proposing a modified firefly algorithm (MFA) with superior performance to the original solution as well as to all the compared metaheuristics. The predictions were performed with the XGBoost technique, which is optimized by the MFA. A synthetic dataset was used for testing, but note that the dataset was generated by a powerful tool that allows for virtually any scenario setup. Finally, to further elaborate on the performance of the proposed method, the authors provide a SHAP analysis.
The main contributions of this work include a robust system for security breach detection for the use case of IoT in healthcare, the proposition of a modified metaheuristic algorithm with improved performance, and a study case on the generated IoT traffic dataset.
  • A robust healthcare security system based on the XGBoost technique optimized by the proposed MFA.
  • Modification proposals for the FA swarm metaheuristic.
  • Improvements validated through extensive testing on a simulated dataset with comparison to eight other XGBoost-metaheuristic optimized solutions.
  • Performance optimization based on the SHAP feature importance which clearly indicates which features contribute to which predicted class.
  • Best performance interpretation using SHAP analysis for better transparency.
The structure of the manuscript is as follows: Section 2 refers to the fundamentals of this research and similar work, Section 3 provide the original algorithm and modifications applied to it, Section 4 describes the used dataset, experimental setup, and employed metrics, Section 5 describes the obtained results, and Section 6 provides a summarization of this work and future work propositions.

2. Background and Preliminaries

2.1. Literature Review

More and more devices IoT are being connected to traditional computer networks. Such networks are commonly referred to as IoT networks. The security of the IoT ecosystem is based on safeguarding sensitive user data, given that IoT devices are finding increasing applications in all spheres of modern life. Data are typically outsourced and stored on remote cloud servers, which are also accessed by IoT devices. As these devices can generate a large amount of sensitive data, it is essential to enhance comprehensive protection [14].
The latest advancements in the field of IoT are being analyzed, and it is necessary to enhance them by utilizing AI and deep learning [15]. It has been emphasized that technological challenges and gaps have been identified in the IoT-based cloud infrastructure, and it is crucial to improve the entire domain [15].
Hathaliya et al. [2] conducted an interesting study, providing an overview of the transformation from Healthcare 1.0 to 4.0. They emphasized that insecure Healthcare 4.0 techniques can lead to privacy breaches of health data. Attackers can gain access to sensitive information such as user email accounts, patient health reports, exchanged messages with relevant parties, and more. Additionally, the authors found that this technology can also achieve success in terms of data exchange efficiency. Thamilarasu et al. [16] were able to achieve high detection accuracy with minimal resource overhead through simulations. They successfully contributed to improving the overall security of IoT devices used by patients by using simulations to mitigate potential attacks.
The scientific paper [17] explores the problem of vulnerability and security in IoT devices, with a focus on the healthcare domain. The authors have determined that, even with a well-organized network infrastructure using traditional approaches, existing protection mechanisms cannot directly guarantee the security of IoT devices against cyber-attacks due to resource limitations and specific IoT protocols. The authors have proposed a new framework for developing a novel solution for contextual security of IoT devices to facilitate the detection of attacks on the network infrastructure. For the purposes of the study, the authors have utilized a newly developed open-source tool called IoT-Flock for generating IoT data. This tool enabled the authors to develop scenarios that can generate normal and malicious traffic. The outcome of the study resulted in the proposal of a new framework for contextual security of IoT devices and highly sensitive scenarios, such as the IoT healthcare environment.
Hussain et al. [18] focused on the impact of IoT on modern life in various domains. Due to limited resources, IoT devices are attractive targets for cyber-attacks. The paper highlights that traditional encryption methods are not sufficient to protect data integrity in IoT networks, so the authors utilized machine learning and deep learning techniques to enhance security. The authors of the paper [19] focused their research on the development of intelligent components for the identification and profiling of IoT devices, as well as the detection of intrusions in complex IoT network environments. For authentication purposes, the authors proposed fingerprinting techniques. For the needs of the study, the authors created a laboratory environment where they collected data based on four states of each device: on, idle, active, and interaction. In addition to this, the authors simulated smart home activities. Two sets of data were also generated: denial of service (DoS) attacks and brute-force attacks on real time streaming protocol (RTSP) protocols. The authors aimed to create a reference IoT dataset that is realistic and can contribute to and enhance future research in terms of training and evaluating intelligent components of IoT systems, where identification, profiling, and potential intrusion detection need to be performed. The research of this paper revolves around the hypothesis that emphasizes the significance of security in healthcare for sustainability, where data security and integrity play a crucial role. Additionally, the datasets and algorithms used in this field are relatively new, and metaheuristics have not been sufficiently tested for application. However, the contribution of this study can be highly significant for the future development of healthcare as a whole.

2.2. Extreme Gradient Boosting

The modern and popular XGBoost algorithm is a powerful and accurate machine learning algorithm [20,21]. To minimize the loss function, it is necessary to optimize the model’s parameters, which have the most significant impact on the algorithm’s behavior. The XGBoost combines several weak models to achieve a model with improved and more precise predictions. This algorithm can achieve extremely high accuracy by utilizing techniques such as gradient boosting, regularization, parameter optimization, and more. It is based on the idea of providing accurate outcome predictions based on learned patterns of complex dependencies between the input data and the target variable.
The XGBoost algorithm, like most modern machine learning (ML) methods, has numerous specific parameters that can be adjusted to achieve optimal results. Depending on the problem at hand, the parameters being optimized and adjusted directly affect the configuration of different model characteristics. When it comes to parameter tuning in the XGBoost, it allows for more precise control during the learning process. The underlying idea is to achieve consistency across factors such as accuracy, speed, and model generalization. To achieve better and more optimal results, an iterative parameter tuning process is sometimes necessary to perform accurate and precise data validation based on previous evaluations [21]. Below is the objective function for the XGBoost [21]:
obj ( Θ ) = L ( θ ) + Ω ( Θ ) ,
The presented equation represents the combined sum of the loss function and regularization. In this equation, Θ represents the set of XGBoost hyperparameters, L ( Θ ) corresponds to the loss function, and  Ω ( Θ ) denotes the regularization term. The regularization term is employed to manage the complexity of the model. Specifically, the loss function is defined as the mean squared error.
L ( Θ ) = i ( y i y i ^ ) 2 ,
The mathematical notation represents y i as the predicted value, while y i ^ represents the predicted value for the target variable, given for each i iteration.
L ( Θ ) = i [ y i ln ( 1 + e y ^ i ) + ( 1 y i ) ln ( 1 + e y ^ i ) ] ,
The function aims to find the difference between observed values, i.e., actual and predicted values. The idea and outcome of the function are to minimize the overall loss function value, resulting in improved classification results.
A correlation can be made between the XGBoost algorithm and Healthcare 4.0 technology [22,23]. It is used to gather, analyze, and predict events, thereby improving the decision-making mechanism. In the mentioned work, there is a connection between the XGBoost algorithm and Healthcare 4.0, where it can be concluded that XGBoost is primarily superior compared to the support vector machine (SVM), random forest (RF), and K-nearest neighbors (KNN) algorithms in predicting the risk of type 2 diabetes. It has been found that XGBoost has a robust generalization mechanism and predictive ability [23]. It can also predict appropriate preventive measures in the future based on a patient’s previous conditions.

2.3. Metaheuristics Approaches and Applications

The challenges of NP-hard nature are common with computing which requires the use of stochastic solutions as deterministic methods are not practical. Different families of metaheuristics are recognized based on the natural phenomena of influence for them, which can be insect behavior or evolution [24,25,26]. The most influential ones are nature-inspired algorithms [27], which are further grouped into genetic and swarm algorithms. The inspiration for such solutions comes from physical phenomena (e.g., gravity, storm), human behavior (e.g., brainstorming, teaching, and learning), and mathematical laws (e.g., trigonometric functions).
Large groups of usually modest units cooperating represent the main inspiration for swarm intelligence algorithms. Birds or insects moving in swarms are capable of manifesting sophisticated behavior patterns that can be translated to algorithms [28,29]. The results of these algorithms indicate their high capability of solving NP-hard solutions for real-world problems. Notable algorithms include ant colony optimization (ACO) [30], particle swarm optimization (PSO) [31], the bat algorithm (BA) [32,33], as well as the FA [34]. A recent group of mathematical-based algorithms proved high-performing as they apply trigonometry to direct the search process like the sine-cosine algorithm (SCA) [35] and the arithmetic optimization algorithm (AOA) [36].
The variety of metaheuristic solutions is a result of the NFL, as new solutions always have to be explored for new problems [37]. No single method can be the optimal solution for all optimization problems. Hence, one algorithm can be the best for one problem but provide the worse results for another. Different approaches are suitable for different applications, and optimization algorithms have been applied to tackling several real-world challenges with some notable recent examples, including COVID-19 case predictions [38] as well as challenges associated with supply and demand in the energy sector [39,40].

2.4. Shapley Additive Explanations

To present the performance of the model clearly, the SHAP method was performed. The method is considered to provide meaningful and straightforward interpretations of the model-derived decision and avoids the trade-off of accuracy and interpretability. The game theory approach is applied to improve individual predictions for feature importance calculation based on Shapley values [41]. With a delegation of difference from prediction to prediction average [42] every cooperating party (feature) receives a joint payout depending on their contribution and the payments distributed in such a manner are considered Shapley values. If a feature takes a baseline value (mean) the SHAP interprets the impact compared to a model’s prediction and gives each feature a measure of importance according to their individual contributions to a specific prediction. By doing so, valuable insights are obtained, the possibility of underestimation of a feature’s importance is minimized, Shapley value generalization-based interaction effects are captured, and the global behavior of the model is interpreted with retention of local faithfulness [43,44].

3. Materials and Methods

The original implementation of the FA is shown in this section, followed by the descriptions of known and observed flaws of the original FA. This section suggests improvements to the original algorithm to address the described flaws.

3.1. Original Firefly Algorithm

The FA [45] is a popular metaheuristic algorithm inspired by the natural behavior of fireflies when searching for their mate. The performs according to the following steps:
  • Initialization,
  • Brightness calculation,
  • Firefly movement,
  • Brightness update,
  • Steps 3 and 4 are repeated until satisfactory convergence or a defined number of iterations is reached.
Initialization is the process of generating a population of fireflies with randomly determined positions within a defined search range. Then, the brightness calculation is performed, where a rule is defined so that each firefly has a specific brightness intensity, which can be calculated using a function [45]. It is important to note that higher function values result in brighter fireflies. The intensity of firefly brightness is determined by the formula:
F i = f ( X i ) ,
where X i represents the position of firefly i, and  f ( X i ) , illustrates the value of the objective function for the corresponding position.
Generally, each firefly can move in the search space toward regions with higher brightness levels—i.e., toward brighter fireflies. The movement of fireflies can be performed using an attraction model based on the correlation between fireflies and their brightness intensity [45]. The idea is that each firefly tends to approach other fireflies with higher light intensity. The movement can be calculated as:
X i ( t + 1 ) = X i ( t ) + β e γ r i j 2 ( X j ( t ) X i ( t ) ) + α ϵ i ( t ) ,
where β 0 represents the attractiveness at r = 0 . However, Equation (5) is commonly swapped for Equation (6) [34]:
β ( r ) = β 0 / 1 + γ × r 2
where X i ( t ) represents the current position of firefly i in the corresponding iteration t, r i j represents the current position of firefly j in the relevant iteration t,
  • β represents the distance between fireflies defined as i and j,
  • β is the attraction factor of fireflies,
  • γ represents the absorption factor of light,
  • α determines the randomness factor,
  • while ϵ i ( t ) represents the random vector.
After the corresponding movements, the light intensity is automatically updated. This updating is performed based on the obtained results from the previously defined fitness function. The concept relies on the idea that if the last (new) value of the function is higher, or better than the previous one, the light intensity is increased; otherwise, it is decreased.
The MFA [46] is a version of the original FA algorithm that incorporates additional modifications aimed at improving performance and convergence. The modifications are often based on specific characteristics of the problem to be optimized. Changes can be observed in the movement of the fireflies, such as a different way of attraction or movement within the search space. The modified algorithm [46] also includes changes in algorithm parameters. These modified parameters can be related to various factors, such as firefly attraction, randomness, or light absorption.
A different strategy can be employed for initializing the population of fireflies, where the parameters influencing the initialization strategy can better cover the search space. Additionally, by adjusting the parameters, a more diverse and improved positioning of fireflies in the search space can be achieved.
The modified algorithm also incorporates mutation operators, which involve various changes in firefly positions that are not solely based on attraction to other fireflies [46]. These mutation operators help avoid local optima. The modified FA algorithm can employ different strategies for selecting the best firefly as a result, which can then be used in the next iteration. In this context, the best result is considered to be the selection of the best individuals.

3.2. Proposed Modified Firefly Algorithm

The FA base version performance is still considered among the top optimizers but the CEC functions [47,48] indicate low performance in certain runs. The FA prioritized less promising regions of the search space. The proposed modifications mitigate the impact of such problems resulting in enhanced exploration.
Upon iteration completion, a fresh solution is made by merger of the current best and a random one from the population. Uniform crossover control parameter shown as p c is used to combine all attributes of both solutions and is determined empirically to p c = 0.1 . Every parameter is subject to mutation regulated by the mutation parameter m p , empirically set to m p = 0.1 . The mutation is performed by selecting a pseudo-random number from range [ l b 2 , u b 2 ] , where the boundaries are represented as u b and l b for upper and lower boundaries, respectively. If the value that will be added or subtracted from the solution parameter is determined by m d , the direction of mutation is another control parameter. Uniform distribution in range [ 0 , 1 ] produces a pseudo-random number ψ , which for values ψ < m d applies subtraction, and otherwise addition. The value of is set to m d = 0.5 .
The worst-performing solution is replaced by the newly generated individual, but the new solution is not evaluated until the next iteration. This results in the modified solution staying equivalent to the elementary version in terms of computation complexity. The new method is called the MFA. The pseudocode of the MFA is given in Algorithm 1.
Algorithm 1 Pseudocode of the suggested MFA
1:
Metaheuristics parameter’s values initial setting
2:
Population P production
3:
P evaluation with regard to the fitness function
4:
for i = 1 to max iteration count do
5:
    for every individual do
6:
       for every better individual do
7:
          if individual is better then
8:
             Obtain attraction in terms of distance
9:
             Adjust position toward the better individual
10:
         end if
11:
      end for
12:
      Evaluate and update individuals in population P
13:
      Produce novel solution by applying genetic crossover mechanism
14:
      Subject novel solution to mutation
15:
      Replace the worst-performing individual with a novel generated solution
16:
    end for
17:
    Return top-performing solution
18:
end for

4. Experiments

4.1. Datasets

The applied IoT healthcare security dataset [49] is a dataset generated by a tool proposed by the authors of the referenced work [17]. This research aims to recreate the experiments from the [17] to establish the basis for comparison. The dataset is publicly available at https://www.kaggle.com/datasets/faisalmalik/iot-healthcare-security-dataset (accessed on 14 July 2023).
The authors [17] applied an open-source IoT traffic generator tool IoT-Flock to create a test case for a healthcare problem. IoT-Flock is a versatile tool for creating simulation environments of normal and malicious devices assigned with IP addresses. It is noteworthy that not many such tools support the attacking capabilities of simulation. This results in traffic containing regular communication as well as attacking patterns that are paramount for better IDS and IPS design. The application supports the generation of messages that adhere to two protocols either message queuing telemetry transport (MQTT) or constrained application protocol (COAP). MQTT-generated devices require device information including the username, password, broker’s IP address, and subscribing or publishing indicators. COAP devices require the IP address of the server and the COAP command. Additionally, XML output alongside a GUI and console mode is supported by the application.
The author [49] dataset is an intensive care unit (ICU) use case that simulates an ICU with two beds each with nine sensors and one control unit labeled according to the scheme: Bedx-Control-unit, where x represents the number of each bed. This unit is responsible for time profile settings, infusion pump dosages, and the emergency alarm that shows if a patient is in need according to sensors. The dataset is created with an additional unit for environment control which sets the temperature and humidity, can detect smoke, and activate the alarm as well. Considering that the data are transmitted by sensors in certain intervals, the authors introduce two additional characteristics, the data and time profiles. Range and data type transmitted are specified by the data counterpart, while the time profile defines the intervals in which the communication between the sensors and the control unit is performed.
The process of data generation for the applied ICU dataset results in two separate networks. The first network represents a model of the described ICU unit, and the second is the network of the attacker. The targeted network consists of MQTT devices that transmit and receive data alongside an MQTT broker. These devices include the patient and environment monitors that communicate under usual circumstances. IoT-Flock creates a virtual network interface for each device on a single physical machine running Linux. The MQTT broker and COAP server were running on a different machine. The network of the invader supports the execution of various attack types on the target network. The attacker network consists of ten devices that can perform four different attack types which are the MQTT distributed denial-of-service, MQTT publish flood, brute force, and SlowITE [50]. More details on the dataset are available in the referenced work [17].
The KNN algorithm with the number of neighbors set to K = 5 was applied for feature selection from the datasets, as the original dataset consists of 50 features in total. The best 10 features were selected for maximization of the proposed methods’ performance and this experiment confirms that the same 10 features are selected as in the referenced work [17]. The features are described in detail in [17] and include:
  • frame.time_delta,
  • tcp.time_delta,
  • tcp.flags.ack,
  • tcp.flags.push,
  • tcp.flags.reset,
  • mqtt.hdrflags,
  • mqtt.msgtype,
  • mqtt.qos,
  • mqtt.retain, and
  • mqtt.ver.
Due to the format of mqtt.hdrflags being hexadecimal, its values had to be converted to decimal format so that the ML method could process them.
The research that is the basis of comparison for this work [17], employs only binary classification for the problem with two types of data. The first is the regular communication without attackers and represents class 0, while the second one represents the attacking case and class 1. The performed experiments are an extension of a previous study from [17] in the form of three classes given below:
  • Class 0—No attack, environment monitoring,
  • Class 1—No attack, patient monitoring, and
  • Class 2—Attack.
Figure 1 displays the binary and multiclass distribution of the applied datasets with pie and bar charts. The heatmap of the features is provided in Figure 2.

4.2. Experimental Setup

Optimization of XGBoost hyperparameters is performed by the proposed MFA algorithm in the case of the ICU dataset. The optimized hyperparameters along their boundaries and variable types are given in the following list:
  • learning rate ( η ), search limits: [ 0.1 , 0.9 ] , continuous variable,
  • m i n _ c h i l d _ w e i g h t , search limits: [ 1 , 10 ] , continuous variable,
  • subsample, search limits: [ 0.01 , 1 ] , continuous variable,
  • collsample_bytree, search limits: [ 0.01 , 1 ] , continuous variable,
  • max_depth, search limits: [ 3 , 10 ] , integer variable and
  • g a m m a , search limits: [ 0 , 0.8 ] , continuous variable.
The provided boundaries were obtained empirically by the trial and error method.
The proposed method is made with Python programming language alongside scipy, numpy, and pandas, standard Python libraries. The XGBoost model was provided by the sci-kit learn package.
The optimized hyperparameter number is given in the following text as l, and it is used to set the length of an array that represents a single encoded solution. The value of l is 6.
The split of the dataset was performed randomly of which 70% is for training, while the other 30% is used for the test set. The data were not normalized.
The goal of the proposed hybrid XGBoost MFA optimized method performance validation is performed by comparing with eight other optimizers including the original FA, for the same problem. The algorithms included in the comparison are the FA [34], genetic algorithm (GA) [51,52], PSO [31], artificial bee colony (ABC) algorithm [53], chimp optimization algorithm (ChOA) [54], differential evolution linear population size reduction constrained optimization with levy flights (COLSHADE) algorithm [55], and self-adapting spherical search (SASS) [56]. The tests consist of every listed algorithm performing the same test with recommended values for control parameters. All solutions have been implemented and independently tested. The same setup was established for all tested solutions with a population of 10 solutions over 10 iterations per run, over 30 runs.

4.3. Performance Metrics

The evaluation of performance was performed as in [17], due to firm comparison grounds. Accuracy, precision, recall, and F1-score are derived from confusion matrixes generated by each compared solution. It is important to emphasize that the accuracy can vary from one scenario to another as the dataset is imbalanced. Standard terminology is applied in descriptions of the measured values. True and false predictions are indicated as true positive (TP) and false positive (FP) for positive predictions, while the negative predictions are shown as false negative (FN) and true negative (TN).
The ratio of correct predictions to missed ones is accuracy [57].
A c c u r a c y = T P + T N T P + T N + F P + F N
Precision represents the ratio of correct predictions to all predicted positives.
Correct predictions and all predictions (wrong and correct) ratio is the precision [57].
P r e c i s i o n = T P F P + T P
Correct predictions and positive cases ratio represent recall [57].
R e c a l l = T P F N + T P
Recall and precision are combined to provide F1-score [57].
F 1 - score = ( 2 × P r e c i s i o n × R e c a l l ) ( P r e c i s i o n + R e c a l l )
Cohen’s kappa coefficient is an agreement measure for two sets of classification while taking into account the classification randomness [58]. Odds of an event happening due to a particular factor [59] compared to the probability of the same event occurring without that factor is the odd ratio calculated by the following equation:
( T P F P ) ( T N F N )
Odds of an event happening due to a particular factor [59] compared to the probability of the same event occurring without that factor is the odd ratio that represents the kappa coefficient [60] calculated by the following equation:
K = P o b s P e x p 1 P e x p
where the observed agreements ( T P + T N ) are represent as P o b s , while the expected agreements [ ( T P + F N ) × ( T P + F P ) + ( F P + T N ) × ( F N + T N ) ] [61] are given as P e x p . The value of K is in the range [ 0 , 1 ] , while lower values indicate less agreement. Different levels of agreement are recognized: no agreement K 0 , slight agreement 0.01–0.20, fair agreement 0.21–0.40, moderate agreement 0.41–0.60, substantial agreement 0.61–0.80, and near to perfect agreement 0.81–1.00 as described in [61].

5. Simulation Results, Comparative Analysis, Validation, and Interpretation

5.1. Experimental Findings and Comparative Analyssis

The Table 1 provides overall results for binary classification for error minimization experiment. The results of precision, recall, and F1-score are provided in Table 2. XG-MFA obtains the highest F1-scores for both classes, macro average, and the weighted average for all other metrics besides the F1-score unlike the previously stated. XG-GA had the best 0 class precision, XG-SSA the best recall for class 0 and macro average, XG-ChOA the best precision for macro average, while the XG-SASS provides the best class 1 precision along the best class 0 recall. Confirming the proposed method’s performance on the binary classification error minimization optimization the XG-MFA obtained the highest results in accuracy terms. The best obtained results are indicating with bold font.
The settings of hyperparameters for the binary classification for error minimization are provided in Table 3.
The Figure 3 displays multiple plots for the binary minimization experiment. The objective convergence and box plot are provided with a swarm plot displaying diversity, along the precision recall curve, receiver operating characteristics curve, and the confusion matrix.
Secondly, the problem of binary classification for Cohen’s kappa coefficient maximization problem is tested due to the imbalanced dataset that is used. Table 4 provides general metrics like Table 1, in which the XG-MFA once again dominates only this time the shortcomings are recorded as the XG-ABC provides the best results for class 0 best, worst, mean, median, and standard deviation. The rest of the results are in favor of XG-MFA.
Furthermore, a similar performance is recorded with kappa maximization metrics as in Table 2, where the proposed method obtains the best results except for precision and recall for class 0, class 1, and macro average. The details are provided in Table 5.
The settings of hyperparameters for the binary classification for kappa maximization are provided in Table 6.
The plots for the Cohen’s kappa experiment are provided in the Figure 4 in the same manner as for the error minimization experiments as in Figure 3.
Table 7 shows the multiclassification error of the metaheuristics methods. XG-MFA performs the best on all performance indicators except for standard deviation and variance. XG-COLSHADE shows the least variation in its performance, as indicated by having the lowest values for both standard deviation and variance. This could be a significant factor when looking for models with consistent results. XG-ABC seems to perform the worst.
Table 8 shows the performance evaluation results using precision, recall, and f1-score metrics for each class (0, 1, 2), as well as the macro average and weighted average across classes. The support row at the bottom indicates the number of instances for each class in the dataset. XG-MFA performs the best in terms of overall precision, recall, and f1-score (weighted average). It also performs best for the precision and f1-score of class 2. XG-FA has the highest recall for class 0 and the highest precision for class 1. However, it does not lead to macro averages or weighted averages. XG-PSO has the highest recall for class 2, indicating it is the best at correctly identifying positive instances of class 2. XG-SSA has the highest precision for class 0 and achieves the highest macro average for precision, which means it performs well at correctly predicting the positive class across the different classes without taking class imbalance into account. The recall of class 1 is the highest for XG-SSA, suggesting that this method is particularly effective in identifying instances of class 1. In general, while some methods may excel in certain metrics, the XG-MFA method appears to be the most balanced, performing very well across all metrics and classes. This suggests that XG-MFA may be the best choice for applications where all classes and metrics are equally important. Also, note that class distribution is imbalanced (Class 0: 9528, Class 1: 23,043, Class 2: 24,038). This might affect the performance of these models, as they could be biased toward the majority classes (classes 1 and 2). The weighted average metric, which takes this imbalance into account, can provide a better overall evaluation of model performance than the macro average.
The combined previous analysis of error metrics and precision, recall, and f1-score reinforces the conclusion that XG-MFA generally performs the best among these algorithms.
Hyperparameter settings are provided for the multiclass classification for error minimization in the Table 9.
Table 10 presents the results of Cohen’s kappa coefficient for different XGBoost-based metaheuristic methods. Cohen’s kappa statistic is a measure of inter-rater reliability. It is used to measure the agreement between two raters who each classify items into mutually exclusive categories. In this case, it is being used to evaluate the performance of these machine-learning models. From this table, it can be seen that XG-MFA has the highest mean and median kappa coefficient, indicating that on average, this method tends to outperform the others in terms of inter-rater reliability. It also has the highest best value, suggesting that in the best case, this method has the highest agreement between predicted and actual labels. On the other hand, XG-SASS has the lowest standard deviation and variance for the kappa coefficient, indicating that the performance of this method is the most consistent. The performance of this method varies the least from the mean performance.
The Figure 5 represents the same plots as Figure 3 and Figure 4 for the multiclass experiment of error minimization, while the Figure 6 does so for the Cohen’s kappa multiclass maximization experiment.
Table 11 presents a breakdown of precision, recall, and f1-score metrics for different XGBoost-based metaheuristic methods for multi-classification tasks. These metrics are calculated for each of the three classes (0, 1, 2) as well as their macro and weighted averages. The XG-MFA method has the highest macro average and weighted average in precision, recall, and f1-score, suggesting it tends to have the best overall performance among the methods. It is also interesting to see the performance of the models on individual classes. For instance, the XG-GA method achieves the best recall for class 0, XG-MFA for class 1, and XG-PSO for class 2.
Table 12 summarizes metaheuristics hyperparameters on which evaluation of Multiclassification Cohen’s kappa coefficient was performed.

5.2. Statistical Validation

In order to further evaluate the experimental outcomes and establish their statistical significance, the highest scores obtained from 30 independent runs of each algorithm under consideration were collected and analyzed as a series of data. The initial step involves determining the appropriate type of statistical tests, whether they should be parametric or non-parametric. Initially, the suitability of parametric tests is assessed by examining the independence, normality, and homoscedasticity of the variance in the data [62]. The independence criterion is satisfied as every individual run of the metaheuristics begins with the generation of a random set of solutions.
The normality assumption was assessed through the use of the Shapiro–Wilk test, with individual problem analysis [63] for all four experiments that were conducted as part of research: binary classification (error and Cohen’s kappa as objectives), and multiclass classification (error and Cohen’s kappa as objectives). The Shapiro–Wilk test was independently applied to each of the algorithms under consideration in all four regarded scenarios, yielding p-values for each. In all instances, the resulting p-values were found to be smaller than 0.05, indicating that the null hypothesis (H0) can be rejected. Therefore, it can be concluded that the observed results are not originating from the normal distribution in all four scenarios. The Shapiro–Wilk test scores for all four scenarios are provided in Table 13.
Since the normality condition was not fulfilled, it is not safe to apply the parametric tests. Consequently, it was proceeded by applying the non-parametric Wilcoxon signed-rank test [64] utilizing the identical data series containing the best values obtained in each run.
The suggested MFA was utilized as the control method, and Wilcoxon signed-rank test has been conducted over the above-mentioned data series. The calculated p-values were in all cases lower than 0.05, for all four observed scenarios. These results clearly suggest that the introduced MFA method performed statistically significantly better in comparison to all contenders for both significance thresholds a l p h a = 0.1 and a l p h a = 0.05 . The outcomes of Wilcoxon singed-rank test have been provided within Table 14.

5.3. Best Models Results Interpretation

SHAP analysis has been recognized as a method for bringing transparency to model performance. Additionally, two different types of SHAP analysis are provided for the performed experiments. First, XGBoost feature importance is established. The importance is determined by considering the weights of the evaluated nodes, and in the case of higher observation dependence on the subject node, the importance is considered higher. The importance is calculated on a single decision tree and averaged over the rest of the trees. Determined feature importance for the best-attained models in both experiments are shown in Figure 7.
Outcomes demonstrated in Figure 7a suggest that based on importance determined by XGBoost, tcp.time_delta plays a high role in determining outcomes in binary classification, followed by the frame.time_detlta, mqtt.hdrflags and mqtt.msgtype features. While additional features are considered they demonstrated lower importance. Similar outcomes are demonstrated for multi-class classification in Figure 7b, with tcp.time_delta showing high importance in determining outcomes. However, relative to the importance of binary classification, the impact is somewhat lower. This feature is similarly followed by frame.time_delta and mqtt.hdrflags features, which now show a higher impact relative to binary classification.
Feature impact is calculated using SHAP similarly through the approximation approach. The process is performed with a given number of samples and the feature importance is calculated locally, on the current set. Analysis has been conducted for both binary and multi-class experiments, and their outcomes are demonstrated in Figure 8.
Feature impact outcomes for binary classification shown in Figure 8a suggest that tcp.time_delta contributes roughly equally to both classes, similarly to frame.time_delta the second highest rated feature. However, outcomes for the mqtt.dfrflages suggest that it plays a more signification role in class 1 classification. The remaining features show a high impact on class 0 decisions. Analysis outcomes for multi-class experiments shown in Figure 8b mirror previous findings, with tcp.time_delta, mqtt.hdrflags, frame.time_delta and tcp.flags.push showing the highest rate of impact. However, class-wise impacts are interesting, indicating that class 1 decisions are mostly impacted by the first two features.
Lastly, nSHAP analysis was performed for the order of 2, 10, and the Shapley–Taylor analysis. This allows us to determine to what extent features, including classification, count toward a certain outcome away from a certain classification being made. In total, three analyses were conducted that utilized nSHAP. Determined n-Shapley values up to the 10th order are shown in Figure 9.
The outcomes of the nSHAP analysis help us observe that interactions between features exist in the higher order for many of the features both for binary and multi-class classification. Additionally, many of the higher-order interactions demonstrate interesting outcomes, often contributing differently to what is experienced for the main contributions.
Table 15 displays the proposed model’s performance compared to the most used machine-learning techniques for similar classification problems. The performance was evaluated by the error minimization problem. The results confirm the performance of the novel solution as it obtains the highest scores for all metrics, except recall where the decision tree obtained the best score.
The attained Shapley–Taylor interaction index outcomes can be observed in Figure 10.
Finally, faithful Shapley interaction index can be seen in Figure 11.

6. Conclusions

The exploration of machine learning algorithms, particularly those optimized by the modified firefly algorithm, has offered a profound perspective on addressing the security challenges associated with the integration of IoT in Healthcare 4.0.
The models developed in this study have demonstrated impressive performance in detecting security issues within IoT devices used in healthcare, emphasizing the efficacy and potential of machine learning coupled with the modified firefly metaheuristics. Notably, XG-MFA emerged as a highly accurate model, exhibiting an accuracy of 0.991733 in multi-classification tasks, underlining its proficiency in addressing complex, real-world security detection challenges.
Further, the application of SHAP enabled a deeper understanding of the factors contributing to security issues in IoT devices. By identifying these key influencers, we can enhance our prevention strategies and design more robust and secure systems, thereby promoting the sustainability of IoT systems in healthcare.
Our results highlight the power and importance of using advanced, optimized machine learning techniques in cybersecurity within the healthcare sector. Nevertheless, despite the promising results achieved, this area still has vast untapped potential. We encourage future work to further explore and enhance the use of machine learning algorithms in ensuring the security of IoT devices in healthcare, especially in light of the rapid development and integration of IoT systems in everyday life.
In conclusion, while challenges persist in the integration of IoT in healthcare, this study proves that strategic use of machine learning algorithms, like the ones optimized by the modified firefly algorithm, can significantly mitigate these issues. This paves the way for sustainable, secure, and more effective healthcare supported by the Internet of Things.

Author Contributions

All authors contributed equally to this work. Conceptualization, M.Z., N.B., B.N. and R.D.; methodology, N.B., N.S. and L.J.; software, N.B., M.Z., L.J. and N.S.; validation, A.T., L.J., N.S. and A.P.; formal analysis, N.S. and B.N.; investigation, N.B., B.N. and A.P.; resources, A.T., A.P. and R.D.; data curation, M.Z., N.S. and A.P.; writing—original draft preparation, N.S., A.T., A.P. and L.J.; writing—review and editing, M.Z., N.B. and R.D.; visualization, N.B., B.N. and N.S.; supervision, M.Z., B.N. and R.D.; project administration, A.T., A.P. and L.J.; funding acquisition, M.Z. and B.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Partial source code accompanied by the csv containing utilized dataset is available via the following Github URL: https://github.com/nbacanin1/Sustainability-Health-4.0-Security (accessed on 14 July 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wehde, M. Healthcare 4.0. IEEE Eng. Manag. Rev. 2019, 47, 24–28. [Google Scholar] [CrossRef]
  2. Hathaliya, J.J.; Tanwar, S. An exhaustive survey on security and privacy issues in Healthcare 4.0. Comput. Commun. 2020, 153, 311–335. [Google Scholar] [CrossRef]
  3. Hathaliya, J.J.; Tanwar, S.; Tyagi, S.; Kumar, N. Securing electronics healthcare records in healthcare 4.0: A biometric-based approach. Comput. Electr. Eng. 2019, 76, 398–410. [Google Scholar] [CrossRef]
  4. Amjad, A.; Kordel, P.; Fernandes, G. A Review on Innovation in Healthcare Sector (Telehealth) through Artificial Intelligence. Sustainability 2023, 15, 6655. [Google Scholar] [CrossRef]
  5. Krishnamoorthy, S.; Dua, A.; Gupta, S. Role of emerging technologies in future IoT-driven Healthcare 4.0 technologies: A survey, current challenges and future directions. J. Ambient. Intell. Humaniz. Comput. 2023, 14, 361–407. [Google Scholar] [CrossRef]
  6. Ali, M.H.; Jaber, M.M.; Abd, S.K.; Rehman, A.; Awan, M.J.; Damaševičius, R.; Bahaj, S.A. Threat Analysis and Distributed Denial of Service (DDoS) Attack Recognition in the Internet of Things (IoT). Electronics 2022, 11, 494. [Google Scholar] [CrossRef]
  7. Padmashree, A.; Krishnamoorthi, M. Decision Tree with Pearson Correlation-based Recursive Feature Elimination Model for Attack Detection in IoT Environment. Inf. Technol. Control 2022, 51, 771–785. [Google Scholar] [CrossRef]
  8. Rana, S.K.; Rana, S.K.; Nisar, K.; Ag Ibrahim, A.A.; Rana, A.K.; Goyal, N.; Chawla, P. Blockchain technology and Artificial Intelligence based decentralized access control model to enable secure interoperability for healthcare. Sustainability 2022, 14, 9471. [Google Scholar] [CrossRef]
  9. Guezzaz, A.; Azrour, M.; Benkirane, S.; Mohy-Eddine, M.; Attou, H.; Douiba, M. A lightweight hybrid intrusion detection framework using machine learning for edge-based IIoT security. Int. Arab. J. Inf. Technol. 2022, 19, 822–830. [Google Scholar] [CrossRef]
  10. Wen, Y.; Liu, L. Comparative Study on Low-Carbon Strategy and Government Subsidy Model of Pharmaceutical Supply Chain. Sustainability 2023, 15, 8345. [Google Scholar] [CrossRef]
  11. Ksibi, A.; Mhamdi, H.; Ayadi, M.; Almuqren, L.; Alqahtani, M.S.; Ansari, M.D.; Sharma, A.; Hedi, S. Secure and Fast Emergency Road Healthcare Service Based on Blockchain Technology for Smart Cities. Sustainability 2023, 15, 5748. [Google Scholar] [CrossRef]
  12. Panagiotou, D.K.; Dounis, A.I. An ANFIS-Fuzzy Tree-GA Model for a Hospital’s Electricity Purchasing Decision-Making Process Integrated with Virtual Cost Concept. Sustainability 2023, 15, 8419. [Google Scholar] [CrossRef]
  13. Joyce, T.; Herrmann, J.M. A review of no free lunch theorems, and their implications for metaheuristic optimisation. Nat.-Inspired Algorithms Appl. Optim. 2018, 744, 27–51. [Google Scholar]
  14. Venckauskas, A.; Stuikys, V.; Damasevicius, R.; Jusas, N. Modelling of Internet of Things units for estimating security-energy-performance relationships for quality of service and environment awareness. Secur. Commun. Netw. 2016, 9, 3324–3339. [Google Scholar] [CrossRef]
  15. Ahmad, W.; Rasool, A.; Javed, A.R.; Baker, T.; Jalil, Z. Cyber security in iot-based cloud computing: A comprehensive survey. Electronics 2022, 11, 16. [Google Scholar] [CrossRef]
  16. Thamilarasu, G.; Odesile, A.; Hoang, A. An intrusion detection system for internet of medical things. IEEE Access 2020, 8, 181560–181576. [Google Scholar] [CrossRef]
  17. Hussain, F.; Abbas, S.G.; Shah, G.A.; Pires, I.M.; Fayyaz, U.U.; Shahzad, F.; Garcia, N.M.; Zdravevski, E. A framework for malicious traffic detection in IoT healthcare environment. Sensors 2021, 21, 3025. [Google Scholar] [CrossRef]
  18. Hussain, F.; Hussain, R.; Hassan, S.A.; Hossain, E. Machine learning in IoT security: Current solutions and future challenges. IEEE Commun. Surv. Tutor. 2020, 22, 1686–1721. [Google Scholar] [CrossRef]
  19. Dadkhah, S.; Mahdikhani, H.; Danso, P.K.; Zohourian, A.; Truong, K.A.; Ghorbani, A.A. Towards the development of a realistic multidimensional IoT profiling dataset. In Proceedings of the 2022 19th Annual International Conference on Privacy, Security & Trust (PST), Fredericton, NB, Canada, 22–24 August 2022; IEEE: New York City, NY, USA, 2022; pp. 1–11. [Google Scholar]
  20. Chen, T.; He, T.; Benesty, M.; Khotilovich, V.; Tang, Y.; Cho, H.; Chen, K.; Mitchell, R.; Cano, I.; Zhou, T.; et al. Xgboost: Extreme gradient boosting. R Package Version 0.4-2 2015, 1, 1–4. [Google Scholar]
  21. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  22. Gupta, A.; Gusain, K.; Popli, B. Verifying the value and veracity of extreme gradient boosted decision trees on a variety of datasets. In Proceedings of the 2016 11th International Conference on Industrial and Information Systems (ICIIS), Roorkee, India, 3–4 December 2016; IEEE: New York City, NY, USA, 2016; pp. 457–462. [Google Scholar]
  23. Wu, Y.; Zhang, Q.; Hu, Y.; Sun-Woo, K.; Zhang, X.; Zhu, H.; Li, S. Novel binary logistic regression model based on feature transformation of XGBoost for type 2 Diabetes Mellitus prediction in healthcare systems. Future Gener. Comput. Syst. 2022, 129, 1–12. [Google Scholar] [CrossRef]
  24. Stegherr, H.; Heider, M.; Hähner, J. Classifying Metaheuristics: Towards a unified multi-level classification system. Nat. Comput. 2020, 21, 155–171. [Google Scholar] [CrossRef]
  25. Emmerich, M.; Shir, O.M.; Wang, H. Evolution strategies. In Handbook of Heuristics; Springer: Berlin/Heidelberg, Germany, 2018; pp. 89–119. [Google Scholar]
  26. Fausto, F.; Reyna-Orta, A.; Cuevas, E.; Andrade, Á.G.; Perez-Cisneros, M. From ants to whales: Metaheuristics for all tastes. Artif. Intell. Rev. 2020, 53, 753–810. [Google Scholar] [CrossRef]
  27. Khurma, R.A.; Aljarah, I.; Sharieh, A.; Elaziz, M.A.; Damaševičius, R.; Krilavičius, T. A Review of the Modification Strategies of the Nature Inspired Algorithms for Feature Selection Problem. Mathematics 2022, 10, 464. [Google Scholar] [CrossRef]
  28. Beni, G. Swarm intelligence. In Complex Social and Behavioral Systems: Game Theory and Agent-Based Models; Springer: New York, NY, USA, 2020; pp. 791–818. [Google Scholar]
  29. Abraham, A.; Guo, H.; Liu, H. Swarm intelligence: Foundations, perspectives and applications. In Swarm Intelligent Systems; Springer: Berlin/Heidelberg, Germany, 2006; pp. 3–25. [Google Scholar]
  30. Dorigo, M.; Birattari, M.; Stutzle, T. Ant colony optimization. IEEE Comput. Intell. Mag. 2006, 1, 28–39. [Google Scholar] [CrossRef]
  31. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; IEEE: New York, NY, USA, 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  32. Yang, X.S. A new metaheuristic bat-inspired algorithm. In Nature Inspired Cooperative Strategies for Optimization (NICSO 2010); Springer: Berlin/Heidelberg, Germany, 2010; pp. 65–74. [Google Scholar]
  33. Yang, X.S.; Gandomi, A.H. Bat algorithm: A novel approach for global engineering optimization. Eng. Comput. 2012, 29, 464–483. [Google Scholar] [CrossRef]
  34. Yang, X.S. Firefly algorithms for multimodal optimization. In Proceedings of the International Symposium on Stochastic Algorithms, Sapporo, Japan, 26–28 October 2009; Springer: Berlin/Heidelberg, Germany, 2009; pp. 169–178. [Google Scholar]
  35. Mirjalili, S. SCA: A sine cosine algorithm for solving optimization problems. Knowl.-Based Syst. 2016, 96, 120–133. [Google Scholar] [CrossRef]
  36. Abualigah, L.; Diabat, A.; Mirjalili, S.; Abd Elaziz, M.; Gandomi, A.H. The arithmetic optimization algorithm. Comput. Methods Appl. Mech. Eng. 2021, 376, 113609. [Google Scholar] [CrossRef]
  37. Wolpert, D.H.; Macready, W.G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef]
  38. Zivkovic, M.; Bacanin, N.; Venkatachalam, K.; Nayyar, A.; Djordjevic, A.; Strumberger, I.; Al-Turjman, F. COVID-19 cases prediction by using hybrid machine learning and beetle antennae search approach. Sustain. Cities Soc. 2021, 66, 102669. [Google Scholar] [CrossRef]
  39. Bacanin, N.; Jovanovic, L.; Zivkovic, M.; Kandasamy, V.; Antonijevic, M.; Deveci, M.; Strumberger, I. Multivariate energy forecasting via metaheuristic tuned long-short term memory and gated recurrent unit neural networks. Inf. Sci. 2023, 642, 119122. [Google Scholar] [CrossRef]
  40. Al-Qaness, M.A.; Ewees, A.A.; Abualigah, L.; AlRassas, A.M.; Thanh, H.V.; Abd Elaziz, M. Evaluating the applications of dendritic neuron model with metaheuristic optimization algorithms for crude-oil-production forecasting. Entropy 2022, 24, 1674. [Google Scholar] [CrossRef] [PubMed]
  41. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  42. Molnar, C. Interpretable Machine Learning; Lulu. Com: Morrisville, NC, USA, 2020. [Google Scholar]
  43. Lundberg, S.M.; Erion, G.; Chen, H.; DeGrave, A.; Prutkin, J.M.; Nair, B.; Katz, R.; Himmelfarb, J.; Bansal, N.; Lee, S.I. From local explanations to global understanding with explainable AI for trees. Nat. Mach. Intell. 2020, 2, 56–67. [Google Scholar] [CrossRef] [PubMed]
  44. Stojić, A.; Stanić, N.; Vuković, G.; Stanišić, S.; Perišić, M.; Šoštarić, A.; Lazić, L. Explainable extreme gradient boosting tree-based prediction of toluene, ethylbenzene and xylene wet deposition. Sci. Total Environ. 2019, 653, 140–147. [Google Scholar] [CrossRef] [PubMed]
  45. Bezdan, T.; Cvetnic, D.; Gajic, L.; Zivkovic, M.; Strumberger, I.; Bacanin, N. Feature Selection by Firefly Algorithm with Improved Initialization Strategy. In Proceedings of the 7th Conference on the Engineering of Computer Based Systems, Novi Sad, Serbia, 26–27 May 2021; pp. 1–8. [Google Scholar]
  46. Bacanin, N.; Zivkovic, M.; Bezdan, T.; Venkatachalam, K.; Abouhawwash, M. Modified firefly algorithm for workflow scheduling in cloud-edge environment. Neural Comput. Appl. 2022, 34, 9043–9068. [Google Scholar] [CrossRef]
  47. Liang, J.J.; Qu, B.; Gong, D.; Yue, C. Problem Definitions and Evaluation Criteria for the CEC 2019 Special Session on Multimodal Multiobjective Optimization; Computational Intelligence Laboratory, Zhengzhou University: Zhengzhou, China, 2019. [Google Scholar]
  48. Liang, J.; Suganthan, P.; Qu, B.; Gong, D.; Yue, C. Problem Definitions and Evaluation Criteria for the CEC 2020 Special Session on Multimodal Multiobjective Optimization; Computational Intelligence Laboratory, Zhengzhou University: Zhengzhou, China, 2019. [Google Scholar] [CrossRef]
  49. Hussain, F. IoT Healthcare Security Dataset. 2023. Available online: https://www.kaggle.com/datasets/faisalmalik/iot-healthcare-security-dataset (accessed on 14 July 2023).
  50. Vaccari, I.; Chiola, G.; Aiello, M.; Mongelli, M.; Cambiaso, E. MQTTset, a new dataset for machine learning techniques on MQTT. Sensors 2020, 20, 6578. [Google Scholar] [CrossRef] [PubMed]
  51. Goldberg, D.E.; Richardson, J. Genetic algorithms with sharing for multimodal function optimization. In Proceedings of the Genetic Algorithms and Their Applications: Proceedings of the Second International Conference on Genetic Algorithms, Hillsdale, NJ, USA, 28–31 July 1987; Lawrence Erlbaum: Hillsdale, NJ, USA, 1987; Volume 4149. [Google Scholar]
  52. Mirjalili, S. Genetic algorithm. In Evolutionary Algorithms and Neural Networks; Springer: Berlin/Heidelberg, Germany, 2019; pp. 43–55. [Google Scholar]
  53. Karaboga, D.; Basturk, B. On the performance of artificial bee colony (ABC) algorithm. Appl. Soft Comput. 2008, 8, 687–697. [Google Scholar] [CrossRef]
  54. Khishe, M.; Mosavi, M.R. Chimp optimization algorithm. Expert Syst. Appl. 2020, 149, 113338. [Google Scholar] [CrossRef]
  55. Gurrola-Ramos, J.; Hernàndez-Aguirre, A.; Dalmau-Cedeño, O. COLSHADE for real-world single-objective constrained optimization problems. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation (CEC), Glasgow, UK, 19–24 July 2020; IEEE: New York, NY, USA, 2020; pp. 1–8. [Google Scholar]
  56. Zhao, J.; Zhang, B.; Guo, X.; Qi, L.; Li, Z. Self-Adapting Spherical Search Algorithm with Differential Evolution for Global Optimization. Mathematics 2022, 10, 4519. [Google Scholar] [CrossRef]
  57. Hossin, M.; Sulaiman, M.N. A review on evaluation metrics for data classification evaluations. Int. J. Data Min. Knowl. Manag. Process. 2015, 5, 1–11. [Google Scholar]
  58. Arabameri, A.; Chen, W.; Lombardo, L.; Blaschke, T.; Tien Bui, D. Hybrid computational intelligence models for improvement gully erosion assessment. Remote Sens. 2020, 12, 140. [Google Scholar] [CrossRef]
  59. Pepe, M.S.; Janes, H.; Longton, G.M.; Leisenring, W.; Newcomb, P. Limitations of the Odds Ratio in Gauging the Performance of a Diagnostic or Prognostic Marker. Am. J. Epidemiol. 2004, 159, 882–890. [Google Scholar] [CrossRef] [PubMed]
  60. Van Den Eeckhaut, M.; Vanwalleghem, T.; Poesen, J.; Govers, G.; Verstraeten, G.; Vandekerckhove, L. Prediction of landslide susceptibility using rare events logistic regression: A case-study in the Flemish Ardennes (Belgium). Geomorphology 2006, 76, 392–410. [Google Scholar] [CrossRef]
  61. Warrens, M.J. Five ways to look at Cohen’s kappa. J. Psychol. Psychother. 2015, 5, 4. [Google Scholar] [CrossRef]
  62. LaTorre, A.; Molina, D.; Osaba, E.; Poyatos, J.; Del Ser, J.; Herrera, F. A prescription of methodological guidelines for comparing bio-inspired optimization algorithms. Swarm Evol. Comput. 2021, 67, 100973. [Google Scholar] [CrossRef]
  63. Shapiro, S.S.; Francia, R. An approximate analysis of variance test for normality. J. Am. Stat. Assoc. 1972, 67, 215–216. [Google Scholar] [CrossRef]
  64. Wilcoxon, F. Individual comparisons by ranking methods. In Breakthroughs in Statistics; Springer: Berlin/Heidelberg, Germany, 1992; pp. 196–202. [Google Scholar]
Figure 1. Binary and multi-class test cases dataset distribution plots.
Figure 1. Binary and multi-class test cases dataset distribution plots.
Sustainability 15 12563 g001
Figure 2. The heatmap of selected features.
Figure 2. The heatmap of selected features.
Sustainability 15 12563 g002
Figure 3. Binary error minimization plots. (a) Objective box plot, (b) Objective convergence, (c) Area under precision recall, (d) Area under receiver operating characteristic, (e) Data diversity, (f) Confusion matrix.
Figure 3. Binary error minimization plots. (a) Objective box plot, (b) Objective convergence, (c) Area under precision recall, (d) Area under receiver operating characteristic, (e) Data diversity, (f) Confusion matrix.
Sustainability 15 12563 g003
Figure 4. Binary Cohen’s kappa coefficient maximization plots. (a) Objective box plot, (b) Objective convergence, (c) Area under precision recall, (d) Area under receiver operating characteristic, (e) Data diversity, (f) Confusion matrix.
Figure 4. Binary Cohen’s kappa coefficient maximization plots. (a) Objective box plot, (b) Objective convergence, (c) Area under precision recall, (d) Area under receiver operating characteristic, (e) Data diversity, (f) Confusion matrix.
Sustainability 15 12563 g004
Figure 5. Multiclass classification error minimization plots. (a) Objective box plot, (b) Objective convergence, (c) Area under precision recall, (d) Area under receiver operating characteristic, (e) Data diversity, (f) Confusion matrix.
Figure 5. Multiclass classification error minimization plots. (a) Objective box plot, (b) Objective convergence, (c) Area under precision recall, (d) Area under receiver operating characteristic, (e) Data diversity, (f) Confusion matrix.
Sustainability 15 12563 g005
Figure 6. Multiclass classification Cohen’s kappa maximization plots. (a) Objective box plot, (b) Objective convergence, (c) Area under precision recall, (d) Area under receiver operating characteristic, (e) Data diversity, (f) Confusion matrix.
Figure 6. Multiclass classification Cohen’s kappa maximization plots. (a) Objective box plot, (b) Objective convergence, (c) Area under precision recall, (d) Area under receiver operating characteristic, (e) Data diversity, (f) Confusion matrix.
Sustainability 15 12563 g006aSustainability 15 12563 g006b
Figure 7. Feature importance. (a) Binary feature importance, (b) Multi-class feature importance.
Figure 7. Feature importance. (a) Binary feature importance, (b) Multi-class feature importance.
Sustainability 15 12563 g007
Figure 8. SHAP analysis. (a) Binary SHAP analysis, (b) Multi-class SHAP analysis.
Figure 8. SHAP analysis. (a) Binary SHAP analysis, (b) Multi-class SHAP analysis.
Sustainability 15 12563 g008
Figure 9. nSHAP analysis of order 10. (a) Binary nSHAP analysis of order 10, (b) Multi-class nSHAP analysis of order 10.
Figure 9. nSHAP analysis of order 10. (a) Binary nSHAP analysis of order 10, (b) Multi-class nSHAP analysis of order 10.
Sustainability 15 12563 g009
Figure 10. Shapley–Taylor analysis. (a) Binary Shapley–Taylor analysis, (b) Multi-class Shapley–Taylor analysis.
Figure 10. Shapley–Taylor analysis. (a) Binary Shapley–Taylor analysis, (b) Multi-class Shapley–Taylor analysis.
Sustainability 15 12563 g010
Figure 11. nSHAP analysis of order 2. (a) Binary nSHAP analysis of order 2, (b) Multi-class nSHAP analysis of order 2.
Figure 11. nSHAP analysis of order 2. (a) Binary nSHAP analysis of order 2, (b) Multi-class nSHAP analysis of order 2.
Sustainability 15 12563 g011
Table 1. Overall results binary classification—error minimization experiment.
Table 1. Overall results binary classification—error minimization experiment.
Classification Error (Objective)Cohen’s Kappa
MethodBestWorstMeanMedianStdVarBestWorstMeanMedianStdVar
XG-MFA0.0030030.0030910.0030360.003030 3.24 × 10 5 1.05 × 10 9 0.9938520.9936720.9937850.993798 6.63 × 10 5 4.40 × 10 9
XG-FA0.0030380.0033030.0031530.003153 7.85 × 10 5 6.16 × 10 9 0.9937800.9932360.9935450.993544 1.61 × 10 4 2.60 × 10 8
XG-GA0.0030380.0031970.0030960.003091 4.82 × 10 5 2.32 × 10 9 0.9937810.9934540.9936620.993671 9.87 × 10 5 9.75 × 10 9
XG-PSO0.0030740.0032150.0031250.003109 4.79 × 10 5 2.30 × 10 9 0.9937080.9934180.9936040.993635 9.82 × 10 5 9.64 × 10 9
XG-ABC0.0031090.0033740.0032130.003171 8.49 × 10 5 7.21 × 10 9 0.9936360.9930920.9934220.993508 1.74 × 10 4 3.03 × 10 8
XG-SSA0.0030380.0032330.0031380.003144 5.99 × 10 5 3.58 × 10 9 0.9937810.9933810.9935760.993563 1.23 × 10 4 1.52 × 10 8
XG-ChOA0.0030380.0031800.0031050.003100 4.21 × 10 5 1.77 × 10 9 0.9937800.9934900.9936440.993653 8.65 × 10 5 7.49 × 10 9
XG-COLSHADE0.0030210.0032150.0031020.003091 5.85 × 10 5 3.43 × 10 9 0.9938170.9934190.9936490.993671 1.19 × 10 4 1.44 × 10 8
XG-SASS0.0030210.0031620.0031050.003118 4.65 × 10 5 2.16 × 10 9 0.9938150.9935260.9936440.993616 9.53 × 10 5 9.09 × 10 9
Table 2. Detailed results per classes for best performing binary classification models—error minimization experiment.
Table 2. Detailed results per classes for best performing binary classification models—error minimization experiment.
MethodMetric01Macro AvgWeighted Avg
XG-MFAprecision0.9965670.9975820.9970740.996998
recall0.9982190.9953410.9967800.996997
f1-score0.9973920.9964600.9969260.996996
XG-FAprecision0.9966280.9974160.9970220.996962
recall0.9980960.9954240.9967600.996962
f1-score0.9973620.9964190.9968900.996961
XG-GAprecision0.9967800.9972080.9969940.996962
recall0.9979430.9956320.9967870.996962
f1-score0.9973610.9964200.9968900.996961
XG-PSOprecision0.9965360.9974570.9969970.996927
recall0.9981270.9952990.9967130.996926
f1-score0.9973310.9963770.9968540.996927
XG-ABCprecision0.9965060.9974150.9969600.996892
recall0.9980960.9952580.9966770.996891
f1-score0.9973000.9963350.9968180.996891
XG-SSAprecision0.9968410.9971250.9969830.996962
recall0.9978820.9957150.9967980.996962
f1-score0.9973610.9964200.9968900.996961
XG-ChOAprecision0.9962930.9978720.9970830.996964
recall0.9984340.9949660.9967000.996962
f1-score0.9973620.9964170.9968900.996961
XG-COLSHADEprecision0.9967190.9973330.9970260.996980
recall0.9980350.9955490.9967920.996979
f1-score0.9973770.9964400.9969080.996979
XG-SASSprecision0.9960200.9982880.9971540.996983
recall0.9987410.9945920.9966670.996979
f1-score0.9973790.9964370.9969080.996979
support32,57124,03856,60956,609
Table 3. Best obtained models’ hyper-parameters binary classification—error minimization experiment.
Table 3. Best obtained models’ hyper-parameters binary classification—error minimization experiment.
MethodLearning RateMin Child WeightSubsampleColsample by TreeMax DepthGamma
XG-MFA0.8268641.7817490.8018240.663691100.120070
XG-FA0.9000001.1289210.7936750.871647100.800000
XG-GA0.7882521.0000001.0000001.000000100.000000
XG-PSO0.6026431.0000000.6804491.000000100.800000
XG-ABC0.9000001.0000000.6737431.00000070.323973
XG-SSA0.7485881.2162291.0000001.000000100.023409
XG-ChOA0.7748251.0000001.0000001.000000100.567155
XG-COLSHADE0.9000002.3244100.8266960.781224100.259289
XG-SASS0.9000001.0000000.9809850.63283280.149104
Table 4. Overall results binary classification—Cohen’s kappa coefficient maximization experiment.
Table 4. Overall results binary classification—Cohen’s kappa coefficient maximization experiment.
Cohen’s Kappa (Objective)Classification Error
MethodBestWorstMeanMedianStdVarBestWorstMeanMedianStdVar
XG-MFA0.9938170.9937070.9937800.993780 3.64 × 10 5 1.33 × 10 9 0.0030210.0030740.0030380.003038 1.77 × 10 5 3.12 × 10 10
XG-FA0.9937460.9933450.9935770.993581 1.36 × 10 4 1.84 × 10 8 0.0030560.0032500.0031380.003136 6.61 × 10 5 4.36 × 10 9
XG-GA0.9937070.9934540.9936210.993635 7.47 × 10 5 5.57 × 10 9 0.0030740.0031970.0031160.003109 3.64 × 10 5 1.32 × 10 9
XG-PSO0.9937440.9935260.9936440.993689 8.44 × 10 5 7.12 × 10 9 0.0030560.0031620.0031050.003083 4.12 × 10 5 1.70 × 10 9
XG-ABC0.9937070.9932000.9934130.993381 1.72 × 10 4 2.97 × 10 8 0.0030740.0033210.0032170.003233 8.40 × 10 5 7.06 × 10 9
XG-SSA0.9937450.9935620.9936900.993707 5.46 × 10 5 2.99 × 10 9 0.0030560.0031440.0030830.003074 2.65 × 10 5 7.02 × 10 10
XG-ChOA0.9937440.9931640.9935850.993617 1.68 × 10 4 2.86 × 10 8 0.0030560.0033390.0031330.003118 8.24 × 10 5 6.78 × 10 9
XG-COLSHADE0.9937800.9931640.9935450.993581 1.91 × 10 4 3.67 × 10 8 0.0030380.0033390.0031530.003136 9.35 × 10 5 8.74 × 10 9
XG-SASS0.9937810.9933450.9936040.993617 1.30 × 10 4 1.69 × 10 8 0.0030380.0032500.0031250.003118 6.33 × 10 5 4.01 × 10 9
Table 5. Detailed results per classes for best performing binary classification models—Cohen’s kappa coefficient maximization experiment.
Table 5. Detailed results per classes for best performing binary classification models—Cohen’s kappa coefficient maximization experiment.
MethodMetric01Macro AvgWeighted Avg
XG-MFAprecision0.9968110.9972080.9970100.996980
recall0.9979430.9956740.9968080.996979
f1-score0.9973760.9964400.9969080.996979
XG-FAprecision0.9967800.9971670.9969730.996944
recall0.9979120.9956320.9967720.996944
f1-score0.9973460.9963990.9968720.996944
XG-GAprecision0.9960800.9980800.9970800.996929
recall0.9985880.9946750.9966310.996926
f1-score0.9973320.9963760.9968530.996926
XG-PSOprecision0.9964750.9975810.9970280.996945
recall0.9982190.9952160.9967180.996944
f1-score0.9973470.9963970.9968720.996943
XG-ABCprecision0.9962620.9978300.9970460.996928
recall0.9984030.9949250.9966640.996926
f1-score0.9973320.9963750.9968540.996926
XG-SSAprecision0.9967800.9971670.9969730.996944
recall0.9979120.9956320.9967720.996944
f1-score0.9973460.9963990.9968720.996944
XG-ChOAprecision0.9965970.9974150.9970060.996945
recall0.9980960.9953820.9967390.996944
f1-score0.9973460.9963980.9968720.996944
XG-COLSHADEprecision0.9963230.9978310.9970770.996963
recall0.9984030.9950080.9967060.996962
f1-score0.9973620.9964170.9968900.996961
XG-SASSprecision0.9967190.9972910.9970050.996962
recall0.9980040.9955490.9967770.996962
f1-score0.9973610.9964190.9968900.996961
support32,57124,03856,60956,609
Table 6. Best obtained models’ hyper-parameters binary classification—Cohen’s kappa coefficient maximization experiment.
Table 6. Best obtained models’ hyper-parameters binary classification—Cohen’s kappa coefficient maximization experiment.
MethodLearning RateMin Child WeightSubsampleColsample by TreeMax DepthGamma
XG-MFA0.9000001.0000000.9939560.790192100.166214
XG-FA0.9000001.0000001.0000001.000000100.000000
XG-GA0.9000001.9290151.0000000.821090100.800000
XG-PSO0.9000001.0000000.9794821.000000100.061248
XG-ABC0.6257761.0000000.6270280.970949100.371569
XG-SSA0.9000001.0000001.0000001.000000100.000000
XG-ChOA0.8762691.5761440.6601260.766469100.800000
XG-COLSHADE0.8746381.5841141.0000001.000000100.000000
XG-SASS0.9000001.2987500.8853881.000000100.358750
Table 7. Overall results multiclass classification—error minimization experiment.
Table 7. Overall results multiclass classification—error minimization experiment.
Classification Error (Objective)Cohen’s Kappa
MethodBestWorstMeanMedianStdVarBestWorstMeanMedianStdVar
XG-MFA0.0082320.0084620.0083220.008320 9.30 × 10 5 8.66 × 10 9 0.9868430.9864750.986699 0.96704 1.49 × 10 4 2.22 × 10 8
XG-FA0.0083200.0087970.0084730.008409 1.58 × 10 4 2.53 × 10 8 0.9867040.9859390.9864580.986560 2.54 × 10 4 6.46 × 10 8
XG-GA0.0083560.0086030.0084460.008435 7.97 × 10 5 6.35 × 10 9 0.9866450.9862510.9864990.986516 1.27 × 10 4 1.62 × 10 8
XG-PSO0.0084090.0087270.0084990.008444 9.73 × 10 5 9.47 × 10 9 0.9865590.9860520.9864150.986502 1.55 × 10 4 2.41 × 10 8
XG-ABC0.0084260.0089920.0086870.008691 1.71 × 10 4 2.93 × 10 8 0.9865320.9856280.9861140.986107 2.74 × 10 4 7.52 × 10 8
XG-SSA0.0083560.0085500.0084460.008426 7.77 × 10 5 6.04 × 10 9 0.9866430.9863340.9865000.986532 1.25 × 10 4 1.56 × 10 8
XG-ChOA0.0083730.0085680.0084640.008462 7.20 × 10 5 5.18 × 10 9 0.9866150.9863050.9864710.986474 1.15 × 10 4 1.33 × 10 8
XG-COLSHADE0.0083030.0084790.0083750.008364 5.55 × 10 5 3.08 × 10 9 0.9867290.9864450.9866120.986629 8.89 × 10 5 7.91 × 10 9
XG-SASS0.0083910.0088500.0085150.008506 1.37 × 10 4 1.90 × 10 8 0.9865880.9858550.9863910.986405 2.20 × 10 4 4.84 × 10 8
Table 8. Detailed results per classes for best performing multiclass classification models—error minimization simulation.
Table 8. Detailed results per classes for best performing multiclass classification models—error minimization simulation.
MethodMetric012Macro AvgWeighted Avg
XG-MFAprecision0.9761500.9920450.9977050.9886330.991773
recall0.9751260.9957470.9945500.9884740.991768
f1-score0.9756380.9938920.9961250.9885520.991768
XG-FAprecision0.9744450.9925980.9976630.9882350.991693
recall0.9764900.9951400.9943840.9886710.991680
f1-score0.9754670.9938670.9960210.9884510.991686
XG-GAprecision0.9755310.9921290.9975800.9884130.991650
recall0.9749160.9956170.9944670.9883330.991644
f1-score0.9752230.9938700.9960210.9883710.991645
XG-PSOprecision0.9759250.9923400.9970820.9884490.991591
recall0.9742860.9951400.9950500.9881580.991591
f1-score0.9751050.9937380.9960650.9883030.991590
XG-ABCprecision0.9755250.9919570.9975800.9883540.991579
recall0.9747060.9954870.9945090.9882340.991574
f1-score0.9751150.9937190.9960420.9882920.991574
XG-SSAprecision0.9768350.9919170.9972470.9886660.991642
recall0.9736570.9958340.9947580.9880830.991644
f1-score0.9752430.9938710.9960010.9883720.991640
XG-ChOAprecision0.9763310.9920860.9972480.9885550.991626
recall0.9740760.9955730.9948000.9881500.991627
f1-score0.9752020.9938270.9960220.9883500.991624
XG-COLSHADEprecision0.9764410.9922570.9972060.9886350.991697
recall0.9743910.9955300.9948830.9882680.991697
f1-score0.9754150.9938910.9960430.9884500.991695
XG-SASSprecision0.9760280.9922570.9971640.9884830.991609
recall0.9742860.9954430.9946000.9881770.991609
f1-score0.9751560.9938470.9959810.9883280.991607
support952823,04324,03856,60956,609
Table 9. Best obtained models’ hyper-parameters multiclass classification—error minimization experiment.
Table 9. Best obtained models’ hyper-parameters multiclass classification—error minimization experiment.
MethodLearning RateMin Child WeightSubsampleColsample by TreeMax DepthGamma
XG-MFA0.5582241.3906461.0000000.754489100.800000
XG-FA0.9000002.3563921.0000001.00000070.800000
XG-GA0.5326081.6712611.0000001.00000090.595526
XG-PSO0.9000001.2099531.0000000.965372100.162178
XG-ABC0.5167261.0000000.7964650.842297100.518501
XG-SSA0.6433811.0000001.0000000.864255100.087284
XG-ChOA0.6503551.0000001.0000000.800611100.000000
XG-COLSHADE0.5782651.9810631.0000000.761847100.010795
XG-SASS0.9000002.0285731.0000001.000000100.800000
Table 10. Overall results multiclass classification—Cohen’s kappa coefficient maximization experiment.
Table 10. Overall results multiclass classification—Cohen’s kappa coefficient maximization experiment.
Cohen’s kappa (Objective)Classification Error
MethodBestWorstMeanMedianStdVarBestWorstMeanMedianStdVar
XG-MFA0.9867850.9864170.9866700.986670 1.25 × 10 4 1.56 × 10 8 0.0082670.0084970.0083390.008338 7.80 × 10 5 6.11 × 10 9
XG-FA0.9866460.9864160.9865850.986589 6.76 × 10 5 4.58 × 10 9 0.0083560.0084970.0083930.008391 4.18 × 10 5 1.75 × 10 9
XG-GA0.9866460.9856300.9863480.986446 2.93 × 10 4 8.62 × 10 8 0.0083560.0089920.0085410.008479 1.84 × 10 4 3.39 × 10 8
XG-PSO0.9865320.9861360.9863690.986405 1.55 × 10 4 2.41 × 10 8 0.0084260.0086740.0085280.008506 9.71 × 10 5 9.42 × 10 9
XG-ABC0.9865020.9860520.9862240.986207 1.54 × 10 4 2.38 × 10 8 0.0084440.0087270.0086180.008629 9.65 × 10 5 9.32 × 10 9
XG-SSA0.9866440.9860780.9864180.986460 1.65 × 10 4 2.75 × 10 8 0.0083560.0087090.0084970.008470 1.04 × 10 4 1.08 × 10 8
XG-ChOA0.9865890.9863030.9864570.986460 8.53 × 10 5 7.27 × 10 9 0.0083910.0085680.0084730.008470 5.29 × 10 5 2.80 × 10 9
XG-COLSHADE0.9866450.9862770.9865110.986546 1.18 × 10 4 1.41 × 10 8 0.0083560.0085850.0084390.008417 7.38 × 10 5 5.44 × 10 9
XG-SASS0.9867020.9865630.9866030.986589 3.96 × 10 5 1.57 × 10 9 0.0083200.0084090.0083820.008391 2.50 × 10 5 6.24 × 10 10
Table 11. Detailed results per classes for best performing multiclass classification models—Cohen’s kappa coefficient maximization simulation.
Table 11. Detailed results per classes for best performing multiclass classification models—Cohen’s kappa coefficient maximization simulation.
MethodMetric012Macro AvgWeighted Avg
XG-MFAprecision0.9764460.9918330.9977040.9886610.991736
recall0.9746010.9960940.9943420.9883460.991733
f1-score0.9755230.9939590.9960200.9885010.991731
XG-FAprecision0.9753330.9922130.9975800.9883750.991651
recall0.9752310.9953570.9945920.9883930.991644
f1-score0.9752820.9937820.9960840.9883830.991646
XG-GAprecision0.9751420.9920850.9977880.9883380.991655
recall0.9757560.9954000.9943420.9884990.991644
f1-score0.9754490.9937400.9960620.9884170.991647
XG-PSOprecision0.9752230.9924670.9972060.9883000.991578
recall0.9749160.9952260.9946750.9882720.991574
f1-score0.9750700.9938460.9959390.9882850.991574
XG-ABCprecision0.9765210.9918300.9972470.9885330.991554
recall0.9734470.9957040.9947580.9879700.991556
f1-score0.9749810.9937630.9960010.9882470.991552
XG-SSAprecision0.9763330.9920860.9972890.9885700.991644
recall0.9741810.9956170.9947580.9881860.991644
f1-score0.9752560.9938490.9960220.9883760.991642
XG-ChOAprecision0.9751260.9920860.9977040.9883050.991617
recall0.9751260.9955730.9943420.9883470.991609
f1-score0.9751260.9938270.9960200.9883240.991611
XG-COLSHADEprecision0.9755350.9920860.9976210.9884140.991651
recall0.9751260.9955730.9944250.9883750.991644
f1-score0.9753310.9938270.9960210.9883930.991645
XG-SASSprecision0.9754360.9921730.9976630.9884240.991687
recall0.9752310.9957040.9943420.9884260.991680
f1-score0.9753330.9939350.9960000.9884230.991681
support952823,04324,03856,60956,609
Table 12. Best obtained models’ hyper-parameters multiclass classification—Cohen’s kappa coefficient maximization experiment.
Table 12. Best obtained models’ hyper-parameters multiclass classification—Cohen’s kappa coefficient maximization experiment.
MethodLearning RateMin Child WeightSubsampleColsample by TreeMax DepthGamma
XG-MFA0.7801521.0000001.0000001.00000090.517589
XG-FA0.9000001.2654821.0000000.98306090.222017
XG-GA0.5063281.0000000.8161450.945690100.800000
XG-PSO0.8371061.0000001.0000001.000000100.559077
XG-ABC0.5665152.0841700.9899901.000000100.569104
XG-SSA0.8972941.1034031.0000001.000000100.612487
XG-ChOA0.9000001.0000001.0000001.00000090.800000
XG-COLSHADE0.4656013.0912521.0000001.00000090.800000
XG-SASS0.9000001.1079561.0000001.00000090.800000
Table 13. Shapiro Wilk normality tests.
Table 13. Shapiro Wilk normality tests.
ProblemXG-MFAXG-FAXG-GAXG-PSOXG-ABCXG-SSAXG-ChOAXG-COLSHADEXG-SASS
Binary Error0.0480.0410.0450.0380.0360.0320.0430.0260.019
Binary Kappa0.0420.0420.0400.0360.0390.0330.0450.0350.026
Multiclass Error0.0210.0300.0420.0310.0430.0250.0190.0200.023
Multiclass Kappa0.0350.0380.0210.0400.0350.0290.0420.0270.030
Table 14. Wilcoxon signed-rank test scores representing p-values for all four scenarios (XG-MFA vs. others).
Table 14. Wilcoxon signed-rank test scores representing p-values for all four scenarios (XG-MFA vs. others).
Problem/p-ValuesXG-FAXG-GAXG-PSOXG-ABCXG-SSAXG-ChOAXG-COLSHADEXG-SASS
Binary Error0.0160.0370.0280.0080.0250.0310.0340.031
Binary Kappa0.0250.0360.0260.0270.0310.0290.0400.039
Multiclass Error0.0340.0350.0300.0190.0350.0350.0410.027
Multiclass Kappa0.0400.0260.0280.0240.0420.0360.0380.033
Table 15. IoT-Flock generated healthcare dataset commonly used machine learning classifier performance for malicious and normal traffic prediction.
Table 15. IoT-Flock generated healthcare dataset commonly used machine learning classifier performance for malicious and normal traffic prediction.
ML ClassifierPrecisionRecallAccuracyF1-Score
NB79.671299.705252.182168.5093
KNN99.650199.686599.487299.5868
RF99.706999.795499.512199.6534
AB99.554799.445799.503799.4748
LogR95.287990.351599.503694.7071
DT99.694599.799299.478899.6389
XG-MFA99.699899.699799.699799.6996
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Savanović, N.; Toskovic, A.; Petrovic, A.; Zivkovic, M.; Damaševičius, R.; Jovanovic, L.; Bacanin, N.; Nikolic, B. Intrusion Detection in Healthcare 4.0 Internet of Things Systems via Metaheuristics Optimized Machine Learning. Sustainability 2023, 15, 12563. https://doi.org/10.3390/su151612563

AMA Style

Savanović N, Toskovic A, Petrovic A, Zivkovic M, Damaševičius R, Jovanovic L, Bacanin N, Nikolic B. Intrusion Detection in Healthcare 4.0 Internet of Things Systems via Metaheuristics Optimized Machine Learning. Sustainability. 2023; 15(16):12563. https://doi.org/10.3390/su151612563

Chicago/Turabian Style

Savanović, Nikola, Ana Toskovic, Aleksandar Petrovic, Miodrag Zivkovic, Robertas Damaševičius, Luka Jovanovic, Nebojsa Bacanin, and Bosko Nikolic. 2023. "Intrusion Detection in Healthcare 4.0 Internet of Things Systems via Metaheuristics Optimized Machine Learning" Sustainability 15, no. 16: 12563. https://doi.org/10.3390/su151612563

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop