Next Article in Journal
How Effective Are Forecasting Models in Predicting Effects of Exoskeletons on Fatigue Progression?
Previous Article in Journal
MRSNet: Multi-Resolution Scale Feature Fusion-Based Universal Density Counting Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Novel Multi-Classification Dynamic Detection Model for Android Malware Based on Improved Zebra Optimization Algorithm and LightGBM

by
Shuncheng Zhou
,
Honghui Li
*,
Xueliang Fu
,
Daoqi Han
and
Xin He
College of Computer and Information Engineering, Inner Mongolia Agricultural University, Hohhot 010018, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(18), 5975; https://doi.org/10.3390/s24185975
Submission received: 26 August 2024 / Revised: 8 September 2024 / Accepted: 13 September 2024 / Published: 14 September 2024
(This article belongs to the Section Intelligent Sensors)

Abstract

:
With the increasing popularity of Android smartphones, malware targeting the Android platform is showing explosive growth. Currently, mainstream detection methods use static analysis methods to extract features of the software and apply machine learning algorithms for detection. However, static analysis methods can be less effective when faced with Android malware that employs sophisticated obfuscation techniques such as altering code structure. In order to effectively detect Android malware and improve the detection accuracy, this paper proposes a dynamic detection model for Android malware based on the combination of an Improved Zebra Optimization Algorithm (IZOA) and Light Gradient Boosting Machine (LightGBM) model, called IZOA-LightGBM. By introducing elite opposition-based learning and firefly perturbation strategies, IZOA enhances the convergence speed and search capability of the traditional zebra optimization algorithm. Then, the IZOA is employed to optimize the LightGBM model hyperparameters for the dynamic detection of Android malware multi-classification. The results from experiments indicate that the overall accuracy of the proposed IZOA-LightGBM model on the CICMalDroid-2020, CCCS-CIC-AndMal-2020, and CIC-AAGM-2017 datasets is 99.75%, 98.86%, and 97.95%, respectively, which are higher than the other comparative models.

1. Introduction

In the intelligent era, the rapid development and diversification of Android applications have led to an increase in Android malware attacks, which have become increasingly severe [1]. Traditional Android malware detection methods based on static analysis, especially signature-based techniques, can effectively identify known Android malware. However, when the code structure and behavioral patterns of Android malware change, static analysis-based detection methods find it difficult to effectively identify their true intent [2]. This leads to a decrease in detection accuracy.
For this reason, researchers have proposed a dynamic analysis method that focuses on the behavior of Android malware [3]. Android malware is detected through the results of its execution in a sandbox environment. To improve the accuracy of Android malware detection, machine learning (ML) algorithms have been widely adopted. Aslan et al. [2] demonstrated that ML methods lead to improved detection of both known and unknown Android malware. Recently, Gradient Boosting Decision Trees (GBDTs) [4] have shown excellent performance in many studies [5]. Among these methods is the Light Gradient Boosting Machine (LightGBM) [6], an improved GBDT with the advantages of faster training and higher efficiency [7]. In a study by Kirubavathi et al. [7], LightGBM achieved an accuracy of 98.05% in Android malware detection.
However, the LightGBM model contains a large number of hyperparameters, and the values of these hyperparameters significantly affect the model’s detection performance. Manual parameter tuning is extremely challenging due to the sheer number of hyperparameters, the model’s inherent complexity, and the interdependencies among these non-linear hyperparameters. In order to construct an efficient LightGBM-based Android malware detection model, this paper proposes using an optimization method to set the values of the LightGBM hyperparameters.
Since hyperparameter optimization problems are usually nonconvex or nondifferentiable optimization problems, traditional optimization methods may obtain local rather than global optimal solutions. In contrast, metaheuristic algorithms can be effective techniques for solving complex, large-search-space, and nonconvex optimization problems belonging to hyperparametric optimization problems [8]. The Zebra Optimization Algorithm (ZOA) is a novel metaheuristic algorithm that has demonstrated better performance compared to traditional metaheuristics [9]. Despite this, the ZOA still suffers from issues of slow convergence and the tendency to fall into local optimal solutions, limiting its performance.
This paper introduces a dynamic analysis detection model for Android malware, utilizing the Improved Zebra Optimization Algorithm (IZOA) and LightGBM, named the IZOA-LightGBM model, aimed at effectively enhancing the accuracy of malware detection. The model improves the initialization process of the ZOA by introducing elite opposition-based learning (EOBL), which expands the search space and accelerates the convergence speed of the algorithm. The population updating process of the ZOA is improved by introducing the firefly disturbance strategy (FDS), which enhances the local search ability and helps the algorithm avoid falling into local optimal solutions. The optimization process of the hyperparameters of the LightGBM model is completed using the IZOA, which ultimately enables the dynamic detection of Android malware. The primary contributions of this study are outlined in the following points:
  • A novel dynamic analysis-based Android malware detection model using the improved ZOA and LightGBM is proposed, called the IZOA-LightGBM model. It effectively improves the accuracy of Android malware detection.
  • The EOBL and FDS are introduced to improve the traditional ZOA and enhance its search capabilities. The improved algorithms are used to optimize the hyperparameters of the LightGBM model.
  • A suite of experiments was designed to explain and validate the proposed model.
The rest part of this paper is organized as follows. Section 2 presents related work in the field of Android malware detection. Section 3 presents the IZOA-LightGBM model proposed in this paper. Section 4 evaluates and analyzes the detection model proposed in this paper. Section 5 summarizes this paper and provides potential directions for future research.

2. Related Work

Android malware detection methods can generally be divided into three categories: static analysis methods, dynamic analysis methods, and hybrid analysis methods. The results of a study by Gorment et al. [10] in 2023 showed that most of the studies used static analyses with 53.3%, followed by dynamic analyses with 28.9%, and hybrid analyses with only 17.8%.
Static analysis methods, also known as code analysis methods, involve examining software code while it is in a non-executing state. This approach gathers static data from the code to determine if the software in question is classified as Android malware. Researchers have proposed various static analysis-based methods for Android malware detection. For instance, Khariwal et al. [11] used information gain to rank permissions and intents with the aim of finding the optimal set of permissions and intents that would enable better detection of Android malware and proposed a new algorithm by applying several ML algorithms such as Support Vector Machine (SVM), Random Forests (RF), Naive Bayes (NB), and K-Nearest Neighbors (KNN) to find the optimal set. The experiments displayed that RF achieved the highest accuracy rate of 94.73%. Dhalaria et al. [12] extracted API call attributes from the classes.dex file, as well as permission and intent attributes from the AndroidManifest.xml file, and then used these attributes for training and testing to classify the application. The results of the study showed that combined features perform better compared to individual features and it was found that the RF and KNN classifiers provided the highest accuracy of 95.9% for classifying Android apps. Shatnawi et al. [13] proposed an Android malware detection method based on Android permissions and API calls. The experimental results showed that the SVM classifier achieved the highest recognition rate among other comparative classifiers. An average accuracy of 94% was achieved with permission features, while 83% was achieved with API call features. However, when complex obfuscation techniques (e.g., code obfuscation, shelling) are used on an application, static analysis might fail to uncover the genuine behavior of the malware.
Dynamic analysis methods, also known as behavioral analysis methods, record behavioral information such as behavioral logs, context parameters, API call sequences, and other behavioral information of Android malware by executing Android malware samples in a virtual controlled environment. Li et al. [14] proposed a novel framework for Android malware detection based on deep learning (DL) and dynamic analysis. The framework first applies embedding and convolutional layers to perform a joint representation of multiple APIs to represent software behavior. Second, the semantic information of each API call is represented using the API’s category, activity, and operation objects. Finally, a bi-directional long and short-term memory network module is used to mine the relational information between APIs. The study’s findings demonstrated that the method achieves an accuracy rate of 97.31%. Chai et al. [15] proposed a dynamic prototype network for small-sample Android malware detection. This network first uses a dynamic convolutional network to extract dynamic features based on sample adaptation. Subsequently, a dual-sample dynamic activation function is introduced to leverage the inter-sample correlation, thereby mitigating the influence of features that lack correlation between samples on the metric learning process. Experiments showed that the accuracy of the method reaches 94.32%. Hwang et al. [16] proposed a two-stage ransomware detection method based on dynamic analysis. The method first uses a Markov model to capture the features of the software, and later uses an RF model for detection. Experiments showed that the accuracy of the method is 97.3%. Mahdavifar et al. [17] proposed a dynamic Android malware detection method using semi-supervised DL. This method is trained using a set of labeled and a set of unlabeled samples and uses dynamic analysis to extract dynamic behavioral profiles as feature vectors. Experimental results showed that the method achieves up to 96.7% accuracy. However, dynamic analysis methods require more computational resources and time because they require malware execution in an isolated environment.
Hybrid analysis methods integrate static analysis and dynamic analysis methods to capitalize on the strengths of each approach. Static analysis methods are first used to examine the Android malware. Then, the Android malware is launched in a simulated environment to observe its actual behavior. Hadiprakoso et al. [18] trained various ML algorithms including SVM, Decision Tree (DT), RF, NB, and KNN to observe their performance on datasets based on both static and dynamic analysis. Experimental results showed that hybrid analysis can improve Android malware detection accuracy by 5% compared to static analysis alone. Ding et al. [19] proposed a model based on hybrid analysis, which first uses a chi-square test for feature selection, followed by RF for classification, and then uses a residual network combined with a long short-term memory network on top of RF classification results for further classification. Experiments showed that the model has an accuracy of 99%. Amer et al. [20] introduced a universal behavioral graph framework to characterize both malicious and benign processes. The behavioral graph model is based on a combination of statistical, contextual, and graph mining attributes that detect both explicit and implicit connections among API functions within a sequence of invocations. Experiments showed that the model achieves 97.7% accuracy. However, the hybrid analysis method might demand additional computational resources and time because it merges static and dynamic analysis and needs to harmonize and integrate the results of the two analysis methods.
In addition, some scholars have suggested Android malware detection methods that utilize improved meta-heuristic algorithms. For instance, Al-Ogaili et al. [21] suggested to improve the population update process of the whale optimization algorithm using weighted arithmetic mean and applied the improved algorithm for feature selection to detect Android malware attacks. Dong et al. [22] proposed a method to improve the initialization process of the firework optimization algorithm using Lévy flights and improved the updating process of the algorithm using an elitist bootstrapping strategy. The improved fireworks optimization algorithm was then applied to optimize the hyperparameters of the SVM model, subsequently, the optimized SVM model was used to detect Android malware. Aldehim et al. [23] proposed an initialization process to improve the black widow optimization algorithm using Gaussian chaotic mapping for feature selection. Liu et al. [24] introduced a probability density function to improve the particle update process of the particle swarm optimization (PSO) [25] algorithm and applied the improved algorithm to prevent malicious attacks. Given that the ZOA is a relatively new optimization algorithm, fewer researchers have contributed improvements. For example, Qi et al. [26] used chaotic mapping to improve the ZOA population initialization process and improved the population updating process of the algorithm using a positive cosine strategy and dynamic adaptive weighting coefficients. DAMA et al. [27] used chaotic sinusoidal mapping to improve the population updating process of the ZOA. Nevertheless, these approaches continue to face issues with slow convergence rates and a propensity for becoming stuck in local optima.

3. Methods

LightGBM, proposed by Ke and colleagues in 2017, is an improvement of the GBDT algorithm [6]. It diminishes computational demands by employing gradient-based one-side sampling, which filters out samples with smaller gradients during training. In addition, the use of the exclusive feature bundling algorithm reduces the number of features used in the model training process and effectively reduces the complexity of the model. The model has received a lot of attention and has been successfully used for many different types of tasks such as classification, regression and ranking [28]. The LightGBM model has the advantages of faster training and higher efficiency [7]; these advantages make it ideal for Android malware detection. However, the performance of LightGBM is heavily influenced by its many hyperparameters. In order to better utilize LightGBM for Android malware detection, this study used the IZOA to optimize the hyperparameters of the LightGBM model. Figure 1 illustrates the structural framework of the method proposed in this paper. Table 1 lists 11 key hyperparameters that require optimization in the LightGBM model [28,29].
The content of this chapter is arranged as follows: Section 3.1 introduces the hyperparameter optimization method for the LightGBM model based on the ZOA; Section 3.2 introduces the process of optimizing hyperparameters with the improved ZOA, namely IZOA; and Section 3.3 offers an in-depth explanation of the IZOA-LightGBM model for the dynamic analysis and detection of Android malware.

3.1. Zebra Optimization Algorithm

The ZOA, introduced by Trojovska et al. [9] in 2022, is a meta-heuristic optimization algorithm. The algorithm solves the optimization problem by simulating the foraging and defensive behaviors of zebras. The ZOA outperforms nine well-known optimization algorithms, such as the Grey Wolf Optimizer (GWO) [30], Genetic Algorithm (GA) [31], PSO [9], etc. The hyperparameter steps for optimizing LightGBM based on the conventional ZOA are as follows:
  • Initialize zebra population: Randomly generate an initial population containing z zebras. Each zebra i is defined as a position vector A i = ( a i , 1 , , a i , j , , a i , 11 ) that corresponds to the 11 hyperparameter values to be optimized in the LightGBM model as described above. a i , j is the j th hyperparameter value, which takes the value interval l j , u j , 1 i z , 1 j 11 . The fitness function F A i is specified as depicted in Equation (1):
    F A i = min 1 A c c u r a c y L i g h t G B M
    where A c c u r a c y L i g h t G B M denotes the accuracy of the LightGBM model.
  • The zebra with the lowest fitness function value in the population is defined as the Pioneer Zebra ( P Z ). The P Z guides the other zebras toward its position during the optimization process.
2.
Foraging phase: In this phase, the j th hyperparameter value of zebra A i is calculated using Equation (2). Subsequently, the position is updated using Equation (3):
a i , j n e w = a i , j + r P Z j φ · a i , j
A i = A i n e w ,         F A i n e w < F A i A i         ,         e l s e
where A i n e w is the new position of zebra A i , a i , j n e w is the updated value of its j th hyperparameter value, and F A i n e w is the updated value of the corresponding fitness function. r represents a random value within the range [ 0 ,   1 ] , P Z j is the j th hyperparameter value of the pioneer zebra, and φ = r o u n d ( 1 + r ) .
3.
Defense phase: In this phase, the zebra either flees upon being attacked by a lion ( M 1 ) or it counterattacks ( M 2 ). Let the selection probability, p [ 0 ,   1 ] , be a random number that determines the strategy. The zebra’s position is then updated based on the chosen strategy, as described by Equation (4). Subsequently, update the zebra position using Equation (3).
a i , j n e w = M 1 : a i , j + β · ( 2 r 1 ) · ( 1 t T ) · a i , j ,         p 0.5 M 2 :   a i , j + r · A Z j φ · a i , j                                 ,         e l s e
where t is the current iteration number, T is the maximum iteration number, and β is a constant value equal to 0.01 [9]. A Z denotes the position of any zebra in the zebra population except the i th one, and A Z j is the value of its j th hyperparameter.
4.
Iteration: In each iteration, the zebra with the lowest fitness value is chosen as the pioneer zebra P Z . Steps 2 and 3 are reiterated until the iteration limit is attained. The P Z selected after the last iteration is considered the optimal solution.

3.2. Improved Zebra Optimization Algorithm

In the ZOA, the initialization population is generated randomly, which can lead to slow convergence [26]. This study improves the initialization process of the ZOA by introducing EOBL, which accelerates the convergence speed and boosts the algorithm’s global search capability. In addition, the ZOA may also fall into local optimal solutions during the population updating process [27]. In this paper, the FDS is introduced to improve the population updating process of the ZOA, further avoiding falling into local optimal solutions and enhancing the local search capability.

3.2.1. Introducing Elite Opposition-Based Learning to Improve Population Initialization

Opposition-based learning (OBL) [32] is a strategy for enhancing ML proposed by Tizhoosh et al. in 2005. Wang et al. introduced the idea of elite learning and proposed EOBL based on OBL [33]. Experimental results demonstrated that the EOBL has better performance than OBL [33].
In this study, elite opposition-based zebras were generated through EOBL. The fitness function values of both the current and elite opposition-based zebras were calculated and compared, and the zebra with the lower value was selected for the subsequent generation. The detailed steps are outlined below:
  • For any zebra A i , its opposition-based zebra is denoted by A i = a i , 1 , , a i , j , , a i , 11 , where a i , j = r · ( l j + u j ) a i , j . If F ( A i ) < F ( A i ) , then A i is defined as an elite zebra, denoted as E i . Otherwise, A i is directly used as a next-generation zebra.
  • All elite zebras in the zebra population are selected to form the elite zebra population. Let e i , j be the j th hyperparameter value of the elite zebra E i ; then, the j th hyperparameter value e i , j * of the elite opposition-based zebra E i * is calculated as shown in Equation (5). Use Equation (6) to select the next generation of zebras.
    e i , j * = r · a j + b j e i , j
    A i = E i *   ,         F E i * < F A i A i   ,         e l s e
    where a j and b j are the lower and upper bounds of the elite zebra population, respectively. a j = m i n [ e 1 , j , , e i , j , , e p , j ] , b j = m a x [ e 1 , j , , e i , j , , e p , j ] , and p denotes the number of elite zebras, which is less than or equal to the total zebra count within the population. F E i * is the value of the fitness function corresponding to the elite opposition-based zebra E i * .
Figure 2 shows the population initialization process based on the EOBL improvement of the ZOA for optimization of the LightGBM hyperparameters.

3.2.2. Introducing Firefly Disturbance Strategy to Update Zebra Positions

The firefly algorithm, introduced by Yang et al. [34] in 2010, is a meta-heuristic algorithm. The idea originates from the attraction and movement between individual fireflies in nature. This paper introduces the FDS, a key strategy in the firefly algorithm, to improve the population update process of the ZOA. Through the FDS, new zebras are generated, the fitness function values of the current zebra and the new zebra are compared, and the lower one is selected as the next-generation zebra.
Following the position adjustment in the defense phase, the position of zebra A i is updated again using the FDS. The j th hyperparameter value of zebra A i is calculated using Equation (7), and the position is updated using Equation (3):
a i , j n e w = a i , j + β · P Z j a i , j + δ · r 1 2
where β = β 0 · e θ S A i , P Z 2 , β 0 , and θ are constants; S A i , P Z is the spatial distance between the current zebra A i and the pioneer zebra P Z ; and δ is the step factor belonging to the interval 0 ,   1 .
Figure 3 illustrates the process of population updating based on the FDS to improve the ZOA.

3.2.3. Detailed Steps to Improve the Zebra Optimization Algorithm

This section describes in detail the complete steps of the IZOA for optimizing the hyperparameters of LightGBM. The flowchart of the IZOA is shown in Figure 4.
The following section describes the specific steps of the IZOA:
  • Setting IZOA parameters: Set the parameters of the IZOA, including the number of zebra population, the minimum and maximum values of hyperparameters to be optimized in the LightGBM model, the maximum number of iterations, and so on.
  • Introducing EOBL to initialize the population: First, randomly generate the zebra population. Then, compare the fitness function values of current zebras and elite opposition-based zebras, selecting the one with the lower value as the next-generation zebras.
  • Update the pioneer zebra: Select the zebra with the lowest fitness function value in the population as the pioneer zebra.
  • Position update during foraging phase: Calculate the new position of the zebra using Equation (2) and update the position using Equation (3). If the new position has a lower fitness function value, update the position; otherwise, retain the original position.
  • Defense phase position update: Calculate the new position of the zebra using Equation (4) based on the value of the random number p . Update the position using Equation (3). If the new position has a lower fitness function value, update the position; otherwise, retain the original position.
  • Introduces the FDS to update the zebra position again: Calculate the new position of the zebra using Equation (7) and update the position using Equation (3). If the new position has a lower fitness function value, update the position; otherwise, retain the original position.
  • Check whether all zebras are traversed: If the current iteration has not traversed all zebras, return to step 4 to update the next zebra’s position. If all zebras have been updated, proceed to the next step.
  • Verify if the iteration limit has been met: If the maximum number of iterations has not been reached, return to step 3 to continue iterating. If reached, output the pioneer zebra of the IZOA, i.e., the optimal hyperparameters of the LightGBM model, and end the optimization process.

3.3. Android Malware Detection Model IZOA-LightGBM

The IZOA-LightGBM model utilizes the IZOA to tune the hyperparameters of LightGBM with the aim of identifying the optimal hyperparameter settings to improve the detection accuracy. Figure 5 illustrates the flowchart of the IZOA-LightGBM model.
The specific steps for constructing the IZOA-LightGBM model are as follows:
  • Begin: Initiate the optimization process and initialize the environment and parameters.
  • Input dataset: Input Android malware dataset based on dynamic analysis.
  • Data preprocessing: Convert the dataset into a format acceptable to the model and perform data preprocessing to enhance model performance.
  • Split the dataset: Divide the dataset into training, validation, and test sets for respective uses in model training, parameter tuning, and performance evaluation.
  • Initialize LightGBM model: Initialize the LightGBM model with default hyperparameters.
  • Optimize the LightGBM model hyperparameters with the IZOA: Use the IZOA to optimize the LightGBM model hyperparameters, evaluate different combinations using the validation set, and pass the selected optimal hyperparameters to the LightGBM model.
  • Obtain the IZOA-LightGBM model: Obtain the IZOA-LightGBM model with optimal hyperparameters.
  • Evaluate the IZOA-LightGBM model: Assess the IZOA-LightGBM model performance with the test set. Calculate the various metrics of the model (e.g., accuracy, recall, F1-score, etc.). If the model meets expectations, proceed to the next step; otherwise, return to step 6 and re-optimize.
  • Output software test results: Output the final model test results and indicators.
  • End: End the testing process, save all necessary results and models.
The pseudo-code of the IZOA-LightGBM model is shown in Algorithm 1.
Algorithm 1: Pseudo-Code of IZOA-LightGBM
  • Start the IZOA-LightGBM model.
  • Input: Dataset.
  • Perform data preprocessing.
  • Initialize the LightGBM model.
  • Define the IZOA fitness function using Equation (1).
  • Set the number of iterations ( T ), the number of zebras’ population ( z ), and the dimension of the problem to be optimized for the IZOA.
  • Use EOBL, i.e., Equations (6) and (7), to create the initial zebra population.
  • Evaluate the fitness values of the zebras.
  • for k = 1   :   T
  •   Update the Pioneer Zebra ( P Z )
  •   for i = 1   :   z
  •    #Phase 1
  •    Calculate the new position of the i th zebra using Equation (2).
  •    Update the position of the i th zebra using Equation (3).
  •    #Phase 2
  •    p = r a n d ( )
  •    if p 0.5 :
  •     Calculate the new position of the i th zebra using M 1 of Equation (4).
  •    else:
  •     Calculate the new position of the i th zebra using M 2 of Equation (4).
  •    end if
  •    Update the position of the i th zebra using Equation (5).
  •    Use the FDS, i.e., Equations (8) and (9), to update the position of the i th zebra again.
  •    Save the P Z .
  •   end for i = 1   :   z
  • end for k = 1   :   T
  • Input the P Z obtained by the IZOA into the LightGBM model.
  • Obtain the LightGBM model with optimal hyperparameters.
  • Evaluate the IZOA-LightGBM model.
  • Use the optimal IZOA-LightGBM model for detection.
  • End the IZOA-LightGBM model.

4. Experiment and Analysis

Through extensive experiments, this study evaluated the performance of the proposed IZOA-LightGBM model. The system’s hardware was equipped with an Intel (R) Core (TM) i7-6700 CPU at 3.40 GHz, with 16.0 GB of RAM. The software setup incorporated Windows 10, Visual Studio Code, and Python 3.11.2. In this section, Section 4.1 describes the dataset used for the evaluation. Section 4.2 details the evaluation metrics. Section 4.3 outlines the preprocessing steps for the dataset. Section 4.4 presents and analyzes the evaluation results of the IZOA-LightGBM model in detail. Section 4.5 summarizes these results.

4.1. Datasets

Three widely recognized public datasets based on dynamic analysis are used to validate the performance of the IZOA-LightGBM model proposed in this paper on different datasets. The first is the Canadian Institute for Cybersecurity Android Malware 2020 (CICMalDroid-2020) dataset [17,35], abbreviated as the CMD dataset for ease of writing. The second is the Canadian Institute for Cybersecurity project in collaboration with the Canadian Centre for Cyber Security Android Malware 2020 (CCCS-CIC-AndMal-2020) dataset [36,37], referred to as the CCA dataset for ease of writing. The third is the Canadian Institute for Cybersecurity Android Adware and General Malware (CIC-AAGM-2017) dataset [38], referred to as the AAGM dataset for ease of writing. These two datasets are ideal for Android malware detection evaluation due to their high quality and diversity [39].

4.1.1. CMD Dataset

The CMD dataset comprises over 17,341 Android samples from various sources, such as AMD, MalDozer, VirusTotal, and the Contagio security blog. For dynamic analysis, the dataset covers five different categories: adware, banking malware, SMS malware, riskware, and benign software. This dataset contains 471 features. Table 2 displays the sample counts for each category.

4.1.2. CCA Dataset

The CCA dataset, which is a publicly available dataset jointly produced by the Canadian Center for Cybersecurity and the Canadian Cybersecurity Institute in 2020. For dynamic analysis, the dataset contains 142 features. There are 14 different types of Android malware categories in this dataset, including adware, backdoor, file infector, no category, Potentially Unwanted Application (PUA), ransomware, riskware, scareware, Trojan, Trojan banker, Trojan dropper, Trojan SMS, Trojan spy, and zero-day vulnerability. However, the authors of the dataset recommend excluding the no category and zero-day vulnerability categories because the data for these categories are incomplete. After removing the no category and zero-day vulnerability categories from the dataset, 12 categories remained. Table 3 displays the sample counts for each category.

4.1.3. AAGM Dataset

The AAGM dataset was compiled by the Canadian Institute for Cyber Security Research. To generate traffic that is representative of the real world, the institute used an Android smartphone (Nexus 5) rather than an emulator or Android virtual device for traffic collection, ensuring both the quality and quantity of the dataset. For dynamic analysis, the dataset contains 80 features. There are three different types of Android malware categories in this dataset: 1500 benign, 250 adware, and 150 general malware. The number of traffic bars for each category is shown in Table 4.

4.2. Evaluation Metrics

The following metrics were utilized to assess the model performance:
  • Training time [17]: The training time is the time taken by the model from the beginning to the end of the training process.
  • Detection time [17]: The detection time is the time taken by the model from receiving the sample to be detected to making a judgment on whether it is malware or not.
  • Model size [17]: The model size is the amount of space occupied by model on the storage medium.
  • Accuracy [40]: Accuracy is the ratio of the correctly identified samples to the total number of samples in the model. The formula is presented in Equation (8):
    A c c u r a c y = T P + T N T P + F P + T N + F N
    where T P (True Positive) represents the number of samples that the model correctly predicted as a category; T N (True Negative) represents the number of samples that the model correctly predicted as other categories; F P (False Positive) represents the number of samples that the model incorrectly predicted as samples from other categories; and F N (False Negative) represents the number of samples that the model incorrectly predicts as other categories for that category.
  • Precision [40]: Precision is the proportion of samples that the model classified as positive that are actually of the positive category. The formula is presented in Equation (9):
    P r e c i s i o n = T P T P + F P
  • Recall rate [40]: The recall rate is the ratio of the number of samples correctly identified as positive by the model to the total number of samples that are actually positive. The formula is presented in Equation (10):
    R e c a l l r a t e = T P T P + F N
  • F1-Score [40]: The F1-score is a reconciled average of the precision and recall and is used to measure the comprehensive performance of the model in the classification task. The formula is presented in Equation (11):
    F 1 S c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l

4.3. Data Preprocessing

This study conducted a series of data preprocessing steps on the CMD and CCA datasets, including dataset balancing, feature selection, and dimensionality reduction. Balancing the dataset made the model predictions fair for all categories, while making the model generalize better for all categories [41]. The feature selection and feature dimensionality reduction were designed to remove irrelevant or redundant features. In doing so, the model complexity is diminished. In most cases, this approach improves the detection accuracy of the model while also reducing the risk of model overfitting [42].

4.3.1. Balancing Datasets

This study employed three oversampling techniques to balance the datasets: the Adaptive Synthetic Sampling Approach (ADASYN) [43], Random Oversampling Approach (ROA) [41], and Synthetic Minority Over-sampling Technique (SMOTE) [44]. For the CMD dataset, the SMS malware category has the largest number of samples (3904), so this study oversampled each category to 4000 samples. For the CCA dataset, the Trojan category has a larger number of samples, totaling 4025. This study oversampled the categories with fewer than 4025 samples to 4025 samples, and undersampled the categories with more than 4025 samples using random undersampling to 4025 samples. The AAGM dataset was balanced to 10,000 entries per category due to the large number of samples and the limited computing power of the experimental equipment. Table 5 shows the accuracy of the datasets generated using different oversampling techniques on the default LightGBM model.
As can be seen from Table 5, for both the CMD, CCA, and AAGM datasets, the oversampling technique generally improves the accuracy of the models, compared to the unbalanced dataset. Both CMD and CCA datasets achieved the highest accuracy using the ROA. The AAGM dataset achieved the highest accuracy using the Adasyn oversampling technique.

4.3.2. Feature Selection

In this study, the balanced CMD, CCA, and AAGM datasets were fed into the LightGBM model with default hyperparameters. The purpose was to determine the importance of features for predicting labels through feature selection. In this study, split gain [6] was used as a measure of feature importance. Using this method, feature columns that have low split gain are identified and retained, and they are combined with label columns to construct a new dataset. The results before and after feature selection are shown in Table 6.
As shown in Table 6, the split gain-based feature selection strategy decreases the feature count of the datasets and marginally improves the model accuracy. This may be due to the fact that the strategy effectively removes redundant features that do not contribute much to the detection results, as well as reduces the noise interference in the data, thus allowing the model to focus more on the information that contributes substantially to the detection results.

4.3.3. Feature Dimensionality Reduction

This study further applied Principal Component Analysis (PCA) [42] to the after feature selection dataset for feature dimensionality reduction. PCA is a method of dimensionality reduction that transforms the high-dimensional space into a low-dimensional one, aiming to preserve the original data structure and important information. Reducing the data dimensions can enable ML models to analyze data more quickly and efficiently. Figure 6, Figure 7 and Figure 8 show the cumulative explained variance for the CMD, CCA, and AAGM datasets, respectively. The cumulative explained variance refers to the proportion of the total variance explained by all principal components that is accounted for by the first n principal components. The results before and after dimensionality reduction are shown in Table 7.
Figure 6 and Table 7 indicate that 99.9% of the cumulative explained variance was achieved with 29 principal components after conducting dimensionality reduction on the CMD dataset. Although the accuracy was reduced by 0.7%, the training time was reduced by 1.5 s. In addition, Figure 7 and Table 7 indicate that 99.9% of the cumulative explained variance was achieved with 108 principal components after dimensionality reduction processing on the CCA dataset. Although the accuracy was reduced by 0.52%, the training time was reduced by 1.2 s. Figure 8 and Table 7 indicate that 99.9% of the cumulative explained variance was achieved with 108 principal components after dimensionality reduction processing on the AAGM dataset. Although the accuracy was reduced by 1.2%, the training time was reduced by 1.5 s.

4.4. Experimental Results and Analysis

To ensure the reliability of the experimental results and to exclude the influence of chance factors, all experiments in this section were repeated 20 times independently in study paper, and the final results were taken as the average of these independent experiments.
When learning dependencies from data, it is crucial to categorize the data into training, validation, and test sets to prevent overfitting [20]. In this study, 80% of the dataset was utilized for model training, 10% was used for validation to adjust the model, and the remaining 10% was allocated for testing and evaluating the model. To further prevent model overfitting, an early stop strategy was used throughout the evaluation process, establishing 500 as the training epoch count, and ending the training when the model accuracy did not improve for 10 consecutive rounds.
During the experiments, some classical optimization algorithms and classifiers were selected as comparison models, including the LightGBM model optimized using the ZOA (ZOA-LightGBM), the LightGBM model optimized using the Simulated Annealing (SA) algorithm (SA-LightGBM), the LightGBM model optimized using the PSO algorithm (PSO-LightGBM), the XGBoost model optimized using the IZOA proposed in this paper (IZOA-XGBoost), the default LightGBM model, Logistic Regression (LR), and a Multi-Layer Perceptron (MLP) with default hyperparameters. The optimal parameters of each optimization algorithm determined after several experiments are shown in Table 8. Both the LightGBM and XGBoost models used the multi-category log loss as their performance metric, suitable for evaluating models in multi-categorization tasks.

4.4.1. Comparison and Analysis of Confusion Matrices

The confusion matrix, as a commonly used performance evaluation method, can intuitively display model detection effects. It is shown for the IZOA-LightGBM model in Figure 9, Figure 10 and Figure 11.
As can be seen in Figure 9, Figure 10 and Figure 11, the majority of the IZOA-LightGBM model samples in this study were focused along the principal diagonal of the confusion matrix. Points on the primary diagonal of the matrix indicate correct classifications, whereas off-diagonal points signify misclassifications. Thus, the high percentage of main diagonal samples is intuitive evidence of the model’s superior detection performance. On the CMD dataset, the IZOA-LightGBM model incorrectly recognized only 5 out of 2000 samples in the test set. This outcome indicates that the model has a strong ability to distinguish between software samples when they are tested.

4.4.2. Accuracy Comparison

Table 9 presents a comparative analysis of the overall accuracy of the eight models on the CMD, CCA, and AAGM datasets, while Table 10 provides a comparison of the accuracy for each individual class.
Table 9 shows that the IZOA-LightGBM model achieves higher overall accuracy on the CMD, CCA, and AAGM datasets, with 99.75%, 98.86%, and 97.95%, respectively. Table 10 indicates that the model also achieves higher accuracy for most categories. This suggests that the IZOA effectively enhances the classification capabilities of the LightGBM model. The accuracy of the LightGBM model combined with other optimization algorithms (ZOA, SA, PSO) is also higher than that of the unoptimized model. This indicates that these algorithms positively influence the model’s performance. However, the LR and MLP models do not perform as well as the decision tree-based detection models on any of the datasets, with particularly lower accuracies on the CMD dataset.

4.4.3. Precision Comparison

Table 11 compares the precision of the aforementioned eight models on the CMD, CCA, and AAGM datasets.
As shown in Table 11, the IZOA-LightGBM model demonstrates superior precision across the CMD, CCA, and AAGM datasets compared to the other models. The ZOA-LightGBM, SA-LightGBM, and PSO-LightGBM models also performed similarly, but lower than the IZOA-LightGBM model in some categories. The IZOA-XGBoost model also performs better overall, but still slightly below the previously mentioned models. The LR and MLP models are less precise than the other models in most categories, probably because of the lack of hyperparameter optimization.

4.4.4. Recall Rate Comparison

Table 12 compares the recall rate of the aforementioned eight models on the CMD, CCA, and AAGM datasets.
As shown in Table 12, the recall rate exhibits a similar trend to that of the precision metric. On the CMD dataset, the IZOA-LightGBM model performs better, with a recall rate at or near 100% for the Benign and Adware classes. Similarly, on the CCA dataset, the FileInfector and Backdoor categories have 100% recall rate. In contrast, unoptimized models like the default LightGBM, LR, and MLP models display inconsistent performance, performing well in some categories but not in others. In addition, for the FileInfector category, most of the models are correctly recognized. On the AAGM dataset, for the Benign category, the recall of the eight models was generally lower than the other two categories. This phenomenon may indicate that the models have some challenges in recognizing the Benign category.

4.4.5. F1-Score Comparison

Table 13 shows the F1-score comparison of the aforementioned eight models on the CMD, CCA, and AAGM datasets.
In contrast to the previous precision and recall rate analyses, the F1-score, as a reconciled average of the two, provides a more comprehensive assessment of model performance. A high F1-score means that the model has a better balance between precision and recall rate. As shown in Table 13, the IZOA-LightGBM model performs better on the CMD and AAGM datasets and in most categories of the CCA dataset. Differences in the F1-scores across datasets can be observed, with the IZOA-LightGBM model showing superior overall performance on the CMD dataset compared to the CCA dataset. This disparity may be attributed to differences in sample distribution, feature representation, etc., which affect the model’s generalization ability. Moreover, experimental results indicate that an increase in the complexity of the sample categories may lead to increased detection difficulty.

4.4.6. Loss Function Convergence Analysis

As shown in Figure 12, Figure 13 and Figure 14, the IZOA-LightGBM model’s loss function variation with the number of iterations for the training, validation, and testing processes on the CMD, CCA, and AAGM datasets is illustrated. The blue line indicates training loss, the orange line indicates validation loss, and the red dots indicate testing loss.
Figure 12, Figure 13 and Figure 14 show that both training and validation losses decrease as the iteration count increases, and then stabilize. The training loss is consistently lower than the validation loss, indicating that the model performs better on the training set than on the validation set. This phenomenon is expected because the model tries to fit the training data as much as possible during the training process. In addition, the gap between the validation loss and the training loss stays within a small range throughout the training process, indicating that the model does not undergo significant overfitting during the training process. The figure indicates that the test loss is approximately the final value of the validation loss, indicating that the model performs similarly on both sets.

4.4.7. Comparison of Different Improvement Strategies

This study verified the necessity of improving the IZOA-LightGBM model. The LightGBM model was optimized using the EOBL-improved ZOA (EZOA), the FDS-improved ZOA (FZOA), and the unimproved ZOA, respectively. The evaluation results are shown in Table 14.
Table 14 indicates that all the different improvement strategies improve the model detection capabilities. Compared with other improvement strategies, the IZOA-LightGBM model incorporates the improvements of both strategies and has better performance.

4.4.8. Time Complexity and Model Size Comparison

The efficiency and deployment feasibility of the model were assessed by comparing the time complexity and model size, respectively. Table 15 presents the evaluation outcomes.
As shown in Table 15, the IZOA-LightGBM model has a higher time complexity, evident in its longer training duration. However, since the training was conducted offline and the resulting model can be reused, the training time is less critical. In addition, the shorter detection time and smaller model size of the IZOA-LightGBM model means that it may be suitable for deployment in devices with limited computing resources for the real-time detection of Android malware.
The LightGBM model optimized with a meta-heuristic algorithm shows improved detection performance on three datasets compared to the default hyperparameter configuration, but with significant increases in training time and model size. The increase in training time may be due to the introduction of the metaheuristic algorithm, and the increase in model size may be due to the larger and deeper trees generated by the LightGBM model. The SA-LightGBM model exhibited the longest training time on both datasets, probably due to the longer optimization time required for the SA. The IZOA-XGBoost model performs consistently on both datasets, but takes longer to train and detect, and it exhibits larger model sizes of 28.43 MB (CMD) and 37.69 MB (CCA), respectively. In addition, the models on the CCA dataset generally exhibit longer training times, which may be due to the fact that the CCA dataset has more sample categories and numbers than the CMD dataset.

4.4.9. Comparison of This Work with Others

This paper reviews some additional relevant studies from the literature review. Most of the studies used ML or DL methods. Mahdavifar et al. [17], the proposers of the CMD dataset, proposed a Pseudo-Labeled Deep Neural Network (PLDNN) detection method and conducted experiments on the CMD dataset. Mohamed et al. [45] analyzed the CMD dataset using four different ML models. Musikawan et al. [40] proposed an effective improved DNN. Ullah et al. [46] proposed an integrated model. Rahali et al. [37] proposed an image-based DNN approach. Jundi et al. [47] proposed a detection model using a combination of XGBoost and Grammar Evaluation (GE). Xie et al. [48] proposed a DL-based model for multiple-label detection. Tang et al. [49] proposed an Android malware classification method based on Hybrid Bytecode Images (HBIs) and a DNN combined with an Attention Mechanism (DNN-AM). Al-Andoli et al. [50] proposed a parallel DL framework called PDL-FEMC. Li et al. [51] proposed SynDroid, a model for detecting Android malware. Alani et al. [52] proposed a malware detection framework called AdStop. Ullah et al. [53] proposed a semantics-driven Federated Learning (FL) approach for Android malware detection. The papers were categorized by dataset and sorted by time of proposal. Table 16 displays the results.
In Table 16, the studies by Musikawan et al. [40], Batouche et al. [56], and Xie et al [48]. are shown to be less accurate due to the inclusion of the two categories that the authors of the CCA dataset recommended for removal. Table 16 illustrates that the IZOA-Light model achieves higher accuracy than the models of Mahdavifar et al. [40] and Padmavathi et al. [55] on the CMD dataset, and of Rahali et al. [37] on the CCA dataset, under conditions where dynamic analysis was utilized and identical categories are detected. A similar trend was maintained on the AAGM dataset.

4.5. Summary

This chapter aimed to evaluate the availability of the IZOA-LightGBM model proposed in this paper in Android malware detection. The CMD, CCA, and AAGM datasets, which are three public datasets, were used to evaluate the detection performance of the IZOA-LightGBM model. First, the CMD and CCA datasets were preprocessed separately, including dataset balancing, feature selection, and feature dimensionality reduction, to improve the accuracy and efficiency during model training and detection. After that, both the IZOA-LightGBM model and the comparative models were evaluated in terms of accuracy, precision, recall rate, F1-score, training time, detection time, and model size. The evaluation results on the CMD, CCA, and AAGM datasets show that the IZOA-LightGBM model exhibits better detection performance compared to LightGBM models combined with other optimization algorithms such as the ZOA, SA, and PSO. In addition, the short detection time and moderate model size of the IZOA-LightGBM model may be suitable for deployment on real-time Android malware detection devices with limited computational resources.

5. Conclusions

Aiming to enhance the accuracy of Android malware detection via dynamic analysis, this paper proposes a novel model named IZOA-LightGBM. This model further improves the convergence speed and search capability of the ZOA by introducing EOBL and the FDS. Through optimizing the hyperparameters of the LightGBM model using the IZOA, the accuracy of the LightGBM model in detecting Android malware was further improved.
The validity of the proposed model was validated by extensive evaluation on three datasets. The evaluation results show that the IZOA-LightGBM model achieved an overall accuracy of 99.75% on the CMD dataset, 98.86% on the CCA dataset, and 97.95% on the AAGM dataset. It outperforms other comparative models, demonstrating the potential of the model for application in Android malware detection.
However, the model complexity needs to be further reduced, and enhancements can still be made to improve detection accuracy. Therefore, upcoming research will investigate additional efficient and compact detection models to enhance accuracy further. This research is preliminary, based on public datasets, and has not been tested in a live setting. Future work can deeply analyze the application of the model in real Android malware detection scenarios, as well as its performance with other types of malware and in different environments, such as malware in a Windows environment. Furthermore, future research will explore additional data sources and a broader range of features to enhance the model’s robustness. The influence of dataset bias on the model performance will also be assessed in future work. In addition, the overfitting of the model needs to be further considered in future work.

Author Contributions

Conceptualization, S.Z., H.L. and X.F.; methodology, S.Z., H.L. and X.F.; software, S.Z. and D.H.; validation, S.Z., H.L. and D.H.; formal analysis, S.Z., H.L. and X.F.; investigation, S.Z., D.H. and X.H.; writing—original draft preparation, S.Z., D.H. and X.H.; writing—review and editing, S.Z., D.H. and X.H.; visualization, S.Z.; supervision, H.L. and X.F.; project administration, H.L.; funding acquisition, X.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant No. 62041211; the Basic Scientific Research Foundation Project of Colleges and Universities directly under the Inner Mongolia Autonomous Region under Grant No. BR22-14-05; the Inner Mongolia Autonomous Region Science and Technology Program under Grant No. 2022YFHH0070; the China Ministry of Education industry–university cooperative education project Grant No. 230806272235841; the Natural Science Foundation project of Inner Mongolia Autonomous Region Grant No. 2024MS06002; and the Inner Mongolia Autonomous Region Graduate Research Innovation Project under Grant.

Data Availability Statement

The CICMalDroid-2020 dataset is available at https://www.unb.ca/cic/datasets/maldroid-2020.html (accessed on 7 May 2024). The CCCS-CIC-AndMal-2020 dataset is available at https://www.unb.ca/cic/datasets/andmal2020.html (accessed on 7 May 2024). The methodology and code of this study, if required, should be obtained by contacting the corresponding author.

Acknowledgments

The authors would like to thank the anonymous reviewers for their constructive comments and suggestions to improve the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, Z. Microsoft Malware Prediction Using LightGBM Model. In Proceedings of the 2022 3rd International Conference on Big Data, Artificial Intelligence and Internet of Things Engineering (ICBAIE), Xi’an, China, 15–17 July 2022; pp. 41–44. [Google Scholar]
  2. Aslan, Ö.A.; Samet, R. A comprehensive review on malware detection approaches. IEEE Access 2020, 8, 6249–6271. [Google Scholar] [CrossRef]
  3. Li, Z.; Feng, X.; Wu, Z.; Yang, C.; Bai, B.; Yang, Q. Classification of atrial fibrillation recurrence based on a convolution neural network with SVM architecture. IEEE Access 2019, 7, 77849–77856. [Google Scholar] [CrossRef]
  4. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  5. Liu, J.; Zhao, Y.; Feng, Y.; Hu, Y.; Ma, X. Semalbert: Semantic-based malware detection with bidirectional encoder representations from transformers. J. Inf. Secur. Appl. 2024, 80, 103690. [Google Scholar] [CrossRef]
  6. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process Syst. 2017, 30, 52. [Google Scholar]
  7. Kirubavathi, G.; Anne, W.R. Behavioral based detection of android ransomware using machine learning techniques. Int. J. Syst. Assur. Eng. 2024, 1–22. [Google Scholar] [CrossRef]
  8. Yang, L.; Shami, A. On hyperparameter optimization of machine learning algorithms: Theory and practice. Neurocomputing 2020, 415, 295–316. [Google Scholar] [CrossRef]
  9. Trojovska, E.; Dehghani, M.; Trojovsky, P. Zebra Optimization Algorithm: A New Bio-Inspired Optimization Algorithm for Solving Optimization Algorithm. IEEE Access 2022, 10, 49445–49473. [Google Scholar] [CrossRef]
  10. Gorment, N.Z.; Selamat, A.; Cheng, L.K.; Krejcar, O. Machine Learning Algorithm for Malware Detection: Taxonomy, Current Challenges, and Future Directions. IEEE Access 2023, 11, 141045–141089. [Google Scholar] [CrossRef]
  11. Khariwal, K.; Singh, J.; Arora, A. IPDroid: Android malware detection using intents and permissions. In Proceedings of the 2020 Fourth World Conference on Smart Trends in Systems, Security and Sustainability (WorldS4), London, UK, 27–28 July 2020; pp. 197–202. [Google Scholar]
  12. Dhalaria, M.; Gandotra, E. A framework for detection of android malware using static features. In Proceedings of the 2020 IEEE 17th India Council International Conference (INDICON), New Delhi, India, 10–13 December 2020; pp. 1–7. [Google Scholar]
  13. Shatnawi, A.S.; Yassen, Q.; Yateem, A. An android malware detection approach based on static feature analysis using machine learning algorithms. Procedia Comput. Sci. 2022, 201, 653–658. [Google Scholar] [CrossRef]
  14. Li, C.; Lv, Q.; Li, N.; Wang, Y.; Sun, D.; Qiao, Y. A novel deep framework for dynamic malware detection based on API sequence intrinsic features. Comput. Secur. 2022, 116, 102686. [Google Scholar] [CrossRef]
  15. Chai, Y.; Du, L.; Qiu, J.; Yin, L.; Tian, Z. Dynamic prototype network based on sample adaptation for few-shot malware detection. IEEE Trans. Knowl. Data Eng. 2022, 35, 4754–4766. [Google Scholar] [CrossRef]
  16. Hwang, J.; Kim, J.; Lee, S.; Kim, K. Two-stage ransomware detection using dynamic analysis and machine learning techniques. Wirel. Pers. Commun. 2020, 112, 2597–2609. [Google Scholar] [CrossRef]
  17. Mahdavifar, S.; Abdul Kadir, A.F.; Fatemi, R.; Alhadidi, D.; Ghorbani, A.A. Dynamic Android Malware Category Classification using Semi-Supervised Deep Learning. In Proceedings of the 2020 IEEE Intl Conf on Dependable, Autonomic and Secure Computing, Intl Conf on Pervasive Intelligence and Computing, Intl Conf on Cloud and Big Data Computing, Intl Conf on Cyber Science and Technology Congress (DASC/PiCom/CBDCom/CyberSciTech), Calgary, AB, Canada, 17–22 August 2020; pp. 515–522. [Google Scholar]
  18. Hadiprakoso, R.B.; Kabetta, H.; Buana, I.K.S. Hybrid-Based Malware Analysis for Effective and Efficiency Android Malware Detection. In Proceedings of the 2020 International Conference on Informatics, Multimedia, Cyber and Information System (ICIMCIS), Jakarta, Indonesia, 19–20 November 2020; pp. 8–12. [Google Scholar]
  19. Ding, C.; Luktarhan, N.; Lu, B.; Zhang, W. A hybrid analysis-based approach to android malware family classification. Entropy 2021, 23, 1009. [Google Scholar] [CrossRef]
  20. Amer, E.; Zelinka, I.; El-Sappagh, S. A multi-perspective malware detection approach through behavioral fusion of api call sequence. Comput. Secur. 2021, 110, 102449. [Google Scholar] [CrossRef]
  21. Al Ogaili, R.R.N.; Alomari, E.S.; Alkorani, M.B.M.; Alyasseri, Z.A.A.; Mohammed, M.A.; Dhanaraj, R.K.; Manickam, S.; Kadry, S.; Anbar, M.; Karuppayah, S. Malware cyberattacks detection using a novel feature selection method based on a modified whale optimization algorithm. Wirel. Netw. 2023, 88, 104–106. [Google Scholar] [CrossRef]
  22. Dong, D.; Ye, Z.; Su, J.; Xie, S.; Cao, Y.; Kochan, R. A malware detection method based on improved fireworks algorithm and support vector machine. In Proceedings of the 2020 IEEE 15th International Conference on Advanced Trends in Radioelectronics, Telecommunications and Computer Engineering (TCSET), Lviv-Slavske, Ukraine, 25–29 February 2020; pp. 846–851. [Google Scholar]
  23. Aldehim, G.; Arasi, M.A.; Khalid, M.; Aljameel, S.S.; Marzouk, R.; Mohsen, H.; Yaseen, I.; Ibrahim, S.S. Gauss-mapping black widow optimization with deep extreme learning machine for android malware classification model. IEEE Access 2023, 11, 87062–87070. [Google Scholar] [CrossRef]
  24. Liu, W.; Yin, F. Anti-malicious Attack of Wireless Sensor Hospital Networks Based on Improved Particle Swarm Optimization Algorithm. J. Test. Eval. 2023, 51, 291–296. [Google Scholar] [CrossRef]
  25. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
  26. Qi, Z.; Peng, S.; Wu, P.; Tseng, M.-L. Renewable Energy Distributed Energy System Optimal Configuration and Performance Analysis: Improved Zebra Optimization Algorithm. Sustainability 2024, 16, 5016. [Google Scholar] [CrossRef]
  27. DAMA, A.; Khalaf, O.I.; Chandra, G.R. Enhancing the Zebra Optimization Algorithm with Chaotic Sinusoidal Map for Versatile Optimization. Iraqi J. Comput. Sci. Math. 2024, 5, 307–319. [Google Scholar] [CrossRef]
  28. Gao, Y.; Hasegawa, H.; Yamaguchi, Y.; Shimada, H. Malware detection using LightGBM with a custom logistic loss function. IEEE Access 2022, 10, 47792–47804. [Google Scholar] [CrossRef]
  29. Taha, A.A.; Malebary, S.J. Hybrid classification of Android malware based on fuzzy clustering and the gradient boosting machine. Neural Comput. Appl. 2021, 33, 6721–6732. [Google Scholar] [CrossRef]
  30. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  31. Holland, J.H. Genetic algorithms. Sci. Am. 1992, 267, 66–73. [Google Scholar] [CrossRef]
  32. Tizhoosh, H.R. Reinforcement learning based on actions and opposite actions. In Proceedings of the International Conference on Artificial Intelligence and Machine Learning, Cairo, Egypt, 10–21 December 2005. [Google Scholar]
  33. Wang, S.; Ding, L.; Xie, C.; Guo, Z.; Hu, Y. Hybrid Differential Evolutionary Algorithms Applying Elite Reverse Learning Strategies. J. Wuhan Univ. J. 2013, 59, 111–116. [Google Scholar]
  34. Yang, X.-S. Firefly algorithm, stochastic test functions and design optimisation. Int. J. Bio-Inspired Comput. 2010, 2, 78–84. [Google Scholar] [CrossRef]
  35. Samaneh, M.; Dima, A.; Ghorbani, A.A. Effective and Efficient Hybrid Android Malware Classification Using Pseudo-Label Stacked Auto-Encoder. J. Netw. Syst. Manag. 2022, 30, 15–30. [Google Scholar]
  36. Keyes, D.S.; Li, B.; Kaur, G.; Lashkari, A.H.; Gagnon, F.; Massicotte, F. EntropLyzer: Android Malware Classification and Characterization Using Entropy Analysis of Dynamic Characteristics. In Proceedings of the 2021 Reconciling Data Analytics, Automation, Privacy, and Security: A Big Data Challenge (RDAAPS), Hamilton, ON, Canada, 18–19 May 2021; pp. 1–12. [Google Scholar]
  37. Rahali, A.; Lashkari, A.H.; Kaur, G.; Taheri, L.; Gagnon, F.; Massicotte, F. DIDroid: Android Malware Classification and Characterization Using Deep Image Learning. In Proceedings of the 2020 the 10th International Conference on Communication and Network Security, Tokyo, Japan, 27–29 November 2020; pp. 70–82. [Google Scholar]
  38. Lashkari, A.H.; Kadir, A.F.A.; Gonzalez, H.; Mbah, K.F.; Ghorbani, A.A. Towards a network-based framework for android malware detection and characterization. In Proceedings of the 2017 15th Annual Conference on Privacy, Security and Trust (PST), Calgary, AB, Canada, 28–30 August 2017; pp. 233–23309. [Google Scholar]
  39. Huang, Y.; Gou, J.; Fan, Z.; Liao, Y.; Zhuang, Y. A multi-label network attack detection approach based on two-stage model fusion. J. Inf. Secur. Appl. 2024, 83, 214–221. [Google Scholar] [CrossRef]
  40. Musikawan, P.; Kongsorot, Y.; You, I.; So-In, C. An Enhanced Deep Learning Neural Network for the Detection and Identification of Android Malware. IEEE Internet Things J. 2023, 10, 8560–8577. [Google Scholar] [CrossRef]
  41. Abdi, L.; Hashemi, S. To combat multi-class imbalanced problems by means of over-sampling techniques. IEEE Trans. Knowl. Data Eng. 2015, 28, 238–251. [Google Scholar] [CrossRef]
  42. Şahin, D.Ö.; Kural, O.E.; Akleylek, S.; Kılıç, E. Permission-based Android malware analysis by using dimension reduction with PCA and LDA. J. Inf. Secur. Appl. 2021, 63, 102995. [Google Scholar] [CrossRef]
  43. He, H.; Bai, Y.; Garcia, E.A.; Li, S. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–8 June 2008; pp. 1322–1328. [Google Scholar]
  44. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  45. Mohamed, S.E.; Ashaf, M.; Ehab, A.; Shereef, O.; Metwaie, H.; Amer, E. Detecting malicious android applications based on API calls and permissions using machine learning algorithms. In Proceedings of the 2021 International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC), Cairo, Egypt, 26–27 May 2021; pp. 1–6. [Google Scholar]
  46. Ullah, F.; Ullah, S.; Naeem, M.R.; Mostarda, L.; Rho, S.; Cheng, X. Cyber-threat detection system using a hybrid approach of transfer learning and multi-model image representation. Sensors 2022, 22, 5883. [Google Scholar] [CrossRef] [PubMed]
  47. Jundi, Z.Z.; Alyasiri, H. Android malware detection based on grammatical evaluation algorithm and xgboost. In Proceedings of the 2023 Al-Sadiq International Conference on Communication and Information Technology (AICCIT), Al-Muthana, Iraq, 4–6 July 2023; pp. 70–75. [Google Scholar]
  48. Xie, J.; Li, S.; Zhang, Y.; Sun, P.; Xu, H. Analysis and detection against network attacks in the overlapping phenomenon of behavior attribute. Comput. Secur. 2022, 121, 102867. [Google Scholar] [CrossRef]
  49. Tang, J.; Xu, W.; Peng, T.; Zhou, S.; Pi, Q.; He, R.; Hu, X. Android malware detection based on a novel mixed bytecode image combined with attention mechanism. J. Inf. Secur. Appl. 2024, 82, 103721. [Google Scholar] [CrossRef]
  50. Al-Andoli, M.N.; Tan, S.C.; Sim, K.S.; Lim, C.P.; Goh, P.Y. Parallel Deep Learning with a hybrid BP-PSO framework for feature extraction and malware classification. Appl. Soft Comput. 2022, 131, 109756. [Google Scholar] [CrossRef]
  51. Li, J.; He, J.; Li, W.; Fang, W.; Yang, G.; Li, T. SynDroid: An adaptive enhanced Android malware classification method based on CTGAN-SVM. Comput. Secur. 2024, 137, 103604. [Google Scholar] [CrossRef]
  52. Alani, M.M.; Awad, A.I. AdStop: Efficient flow-based mobile adware detection using machine learning. Comput. Secur. 2022, 117, 102718. [Google Scholar] [CrossRef]
  53. Ullah, F.; Cacciagrano, D.; Chen, C.-M.; Kumari, S. Semantic-based Federated Defense for Distributed Malicious Attacks. IEEE Consum. Electron. Mag. 2024, 2162–2248. [Google Scholar] [CrossRef]
  54. Ahmed, M.F.; Biash, Z.T.; Shakil, A.R.; Ryen, A.A.N.; Hossain, A.; Ashraf, F.B.; Hossain, M.I. Shieldroid: A hybrid approach integrating machine and deep learning for android malware detection. In Proceedings of the 2022 International Conference on Decision Aid Sciences and Applications (DASA), Chiangrai, Thailand, 23–25 March 2022; pp. 911–916. [Google Scholar]
  55. Padmavathi, G.; Shanmugapriya, D.; Roshni, A. Performance analysis of unsupervised machine learning methods for mobile malware detection. In Proceedings of the 2022 9th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 23–25 March 2022; pp. 201–206. [Google Scholar]
  56. Batouche, A.; Jahankhani, H. A comprehensive approach to android malware detection using machine learning. In Information Security Technologies for Controlling Pandemics; Springer: Cham, Switzerland, 2021; pp. 171–212. [Google Scholar]
  57. Islam, R.; Sayed, M.I.; Saha, S.; Hossain, M.J.; Masud, M.A. Android malware classification using optimum feature selection and ensemble machine learning. Internet Things Cyber-Phys. Syst. 2023, 3, 100–111. [Google Scholar] [CrossRef]
  58. Bovenzi, G.; Cerasuolo, F.; Montieri, A.; Nascita, A.; Persico, V.; Pescapé, A. A comparison of machine and deep learning models for detection and classification of android malware traffic. In Proceedings of the 2022 IEEE Symposium on Computers and Communications (ISCC), Rhodes, Greece, 30 June–3 July 2022; pp. 1–6. [Google Scholar]
Figure 1. Structural framework of IZOA-LightGBM model.
Figure 1. Structural framework of IZOA-LightGBM model.
Sensors 24 05975 g001
Figure 2. Flowchart of the introduction of EOBL for improved population initialization.
Figure 2. Flowchart of the introduction of EOBL for improved population initialization.
Sensors 24 05975 g002
Figure 3. Flowchart of the introduction of FDS to improve zebra position updating.
Figure 3. Flowchart of the introduction of FDS to improve zebra position updating.
Sensors 24 05975 g003
Figure 4. Flowchart of the IZOA.
Figure 4. Flowchart of the IZOA.
Sensors 24 05975 g004
Figure 5. Flowchart of the IZOA-LightGBM model.
Figure 5. Flowchart of the IZOA-LightGBM model.
Sensors 24 05975 g005
Figure 6. Variance distribution of the number of principal components of the CMD dataset.
Figure 6. Variance distribution of the number of principal components of the CMD dataset.
Sensors 24 05975 g006
Figure 7. Variance distribution of the number of principal components of the CCA dataset.
Figure 7. Variance distribution of the number of principal components of the CCA dataset.
Sensors 24 05975 g007
Figure 8. Variance distribution of the number of principal components of the AAGM dataset.
Figure 8. Variance distribution of the number of principal components of the AAGM dataset.
Sensors 24 05975 g008
Figure 9. Confusion matrix of IZOA-LightGBM on CMD.
Figure 9. Confusion matrix of IZOA-LightGBM on CMD.
Sensors 24 05975 g009
Figure 10. Confusion matrix of IZOA-LightGBM on CCA.
Figure 10. Confusion matrix of IZOA-LightGBM on CCA.
Sensors 24 05975 g010
Figure 11. Confusion matrix of IZOA-LightGBM on AAGM.
Figure 11. Confusion matrix of IZOA-LightGBM on AAGM.
Sensors 24 05975 g011
Figure 12. Loss function curve for CMD dataset.
Figure 12. Loss function curve for CMD dataset.
Sensors 24 05975 g012
Figure 13. Loss function curve for CCA dataset.
Figure 13. Loss function curve for CCA dataset.
Sensors 24 05975 g013
Figure 14. Loss function curve for AAGM dataset.
Figure 14. Loss function curve for AAGM dataset.
Sensors 24 05975 g014
Table 1. Hyperparameters for LightGBM model optimization.
Table 1. Hyperparameters for LightGBM model optimization.
HyperparametersMeaning
learning_rateLearning rate
min_child_samplesMinimum number of samples required for leaf nodes
max_depthMaximum depth
num_leavesNumber of leaf nodes in each tree
max_binMaximum possible number of eigenvalue bins
min_data_in_leafMinimum number of samples for leaf nodes
feature_fractionProportion of feature subsets used to train the model
bagging_fractionControl the sampling ratio of the training data
bagging_freqFrequency of bagging
reg_alphaL1 regular term coefficient
reg_lambdaL2 regular term coefficient
Table 2. CMD dataset.
Table 2. CMD dataset.
NameCount
Adware1253
Banking malware2100
SMS malware3904
Riskware2546
Benign software1795
Total11,598
Table 3. CCA dataset.
Table 3. CCA dataset.
NameCount
Adware5142
Backdoor546
FileInfector119
PUA625
Ransomwore1550
Riskware6792
Scareware424
Trojan4025
Trojan_Banker123
Trojan_Dropper733
Trojan_SMS911
Trojan_Spy1039
Total22,029
Table 4. AAGM dataset.
Table 4. AAGM dataset.
NameCount
Benign471,597
Adware155,613
GenMalware4745
Total631,955
Table 5. Comparison of different oversampling techniques.
Table 5. Comparison of different oversampling techniques.
DatasetsTechniquesAccuracy (%)
CMDROA98.4
ADASYN97.8
SMOTE97.3
Unbalanced93.8
CCAROA96.33
ADASYN95.63
SMOTE95.21
Unbalanced85.51
AAGMROA90.36
ADASYN91.45
SMOTE90.09
Unbalanced88.53
Bold indicates the highest accuracy.
Table 6. Feature selection results.
Table 6. Feature selection results.
DatasetsOriginal Feature CountRemaining Feature CountAccuracy before Feature Selection (%)Accuracy after Feature Selection (%)
CMD47123098.498.5
CCA14211596.3396.46
AAGM807191.4591.77
Bold indicates the highest accuracy.
Table 7. Comparison of dataset before and after dimensionality reduction.
Table 7. Comparison of dataset before and after dimensionality reduction.
Datasets Feature CountAccuracy (%)Training Time (s)
CMDBefore Dimensionality Reduction23098.54.3
After Dimensionality Reduction2997.81.0
CCABefore Dimensionality Reduction11596.4611.7
After Dimensionality Reduction10895.9410.5
AAGMBefore Dimensionality Reduction8091.775.6
After Dimensionality Reduction4690.574.1
Table 8. Parameters of optimization algorithms.
Table 8. Parameters of optimization algorithms.
AlgorithmsParametersValueMeaning
IZOApop30Population number
maxiter100Maximum number of iterations
dim11Problem dimension
θ 2Controls the amplitude of disturbance
β 0 2Initial attraction coefficient
δ 1Rate of attraction decay adjustment
ZOApop30Population number
maxiter100Maximum number of iterations
dim11Problem dimension
SAmaxiter100Maximum number of iterations
dim11Problem dimension
tolfun1 × 10−9Tolerance for change
PSOps30Number of particles
maxiter100Maximum number of iterations
dim11Problem dimension
w9Inertia factor
c12Acceleration constant
c22Acceleration constant
v_min−5Lower bound of velocity
v_max5Lower bound of velocity
Table 9. Comparison of overall accuracy.
Table 9. Comparison of overall accuracy.
ModelAccuracy (%)
CMDCCAAAGM
IZOA-LightGBM99.7598.8697.95
ZOA-LightGBM98.7697.7696.44
SA-LightGBM98.8597.9396.41
PSO-LightGBM98.9497.5497.04
IZOA-XGBoost98.4097.4796.97
LightGBM97.895.9490.57
LR46.5773.3360.97
MLP55.5591.8479.33
Bold indicates the highest accuracy.
Table 10. Comparison of each category’s accuracy.
Table 10. Comparison of each category’s accuracy.
DatasetsCategoryModel
IZOA-LightGBMZOA-LightGBMSA-LightGBMPSO-LightGBMIZOA-XGBoostLightGBMLRMLP
CMDBenign10099.2799.5199.2799.5196.3559.3769.59
Adware99.7698.3298.569998.3298.0820.6267.63
Bank Malware99.7398.9499.299.298.9397.3418.0423.87
SMS Malware99.7599.2498.9899.2497.4797.7291.6734.6
Riskware99.597.9997.9997.9997.7493.9842.3679.2
CCAAdware97.6595.0496.696.693.4788..2553.273.15
Backdoor10099.7599.7510099.7599.7570.899.74
FileInfector10010099.7610099.7699.7696.21100
PUA99.7499.7499.4899.4899.7399.7380.6199.05
Ransomware96.6594.2695.796.5394.2694.2676.3480.67
Riskware98.1294.694.693.393.986.8561.3779.46
Scareware99.2610099.59910099.584.2499.26
Trojan98.1392.2992.2892.2991.128558.0582.06
Trojan_Banker99.299.7399.7399.1999.7399.1988.6898.54
Trojan_Dropper99.7599.7499.7499.7499.7299.1958.3798.14
Trojan_SMS98.4999.2498.7498.4899.2399.2472.6694.27
Trojan_Spy99.599.4999.4998.59998.580.0995.65
AAGMGenMalware98.3197.6897.2998.0698.2691.9666.3782.12
Adware98.1196.6296.5197.2197.0193.0363.9882.19
Benign97.4194.9895.3895.7895.9886.6550.170.68
Bold indicates the highest accuracy.
Table 11. Precision comparison.
Table 11. Precision comparison.
DatasetsCategoryModel
IZOA-LightGBMZOA-LightGBMSA-LightGBMPSO-LightGBMIZOA-XGBoostLightGBMLRMLP
CMDBenign99.7698.7999.0398.7999.0396.1977.4674.67
Adware99.7698.898.898.898.0997.5653.7565.89
Bank Malware99.7398.4298.6898.9497.3998.3372.3479.65
SMS Malware10098.598.4998.9998.4799.7431.6282.53
Riskware99.599.2499.2499.2498.9896.959.7234.73
CCAAdware99.7397.5997.6397.6397.5592.164.8686.59
Backdoor98.0596.8796.8896.1896.8797.1171.9291.25
FileInfector97.8799.2899.0497.8799.2899.0481.4199.75
PUA99.7498.7198.9697.9497.295.2673.1895.01
Ransomware98.0697.0496.6296.5397.0497.0465.9394.35
Riskware10099.0299.0299.0298.5292.0474.0486.9
Scareware99.0198.7898.5398.5298.7898.7886.1597.58
Trojan99.2997.2997.2997.2996.5391.4671.6683.83
Trojan_Banker99.299.7399.7398.9399.4799.4774.0799.22
Trojan_Dropper98.0397.5597.3196.8497.5597.5572.7587.01
Trojan_SMS98.4994.0395.3894.994.0394.0361.8688.29
Trojan_Spy99.0197.5699.019997.0997.0980.8392.07
AAGMGenMalware98.0396.019696.4896.491.0770.7582.12
Adware97.5396.3396.3397.1297.3189.353.1873.94
Benign98.2997.0396.9497.5597.2591.4260.8583.01
Bold indicates the highest precision.
Table 12. Recall rate comparison.
Table 12. Recall rate comparison.
DatasetsCategoryModel
IZOA-LightGBMZOA-LightGBMSA-LightGBMPSO-LightGBMIZOA-XGBoostLightGBMLRMLP
CMDBenign10099.2799.5199.2799.5196.7559.3769.59
Adware99.7698.3298.5699.0498.3298.0820.6267.63
Bank Malware99.7398.9499.299.298.9497.6518.0423.87
SMS Malware99.7599.2498.9999.2497.4797.7391.6734.6
Riskware99.597.9997.9997.9997.7495.9842.3679.2
CCAAdware97.6595.0496.6196.6193.4788.2553.273.15
Backdoor10099.7510010099.7510070.899.74
FileInfector10010010010010010096.21100
PUA99.7599.7499.4899.4899.7499.7480.6199.05
Ransomware96.6594.2695.6993.394.2694.2676.3480.66
Riskware98.1294.694.694.693.986.8561.3779.46
Scareware99.2610099.599.0110010084.2499.26
Trojan98.1392.2992.2992.2991.1285.0558.0582.06
Trojan_Banker99.299.7399.7399.299.7310088.68100
Trojan_Dropper99.7799.7599.7599.7599.7599.7558.3798.14
Trojan_SMS98.4999.2498.7498.4999.2599.2472.6694.27
Trojan_Spy99.699.599.598.5199.599.580.0995.65
AAGMGenMalware98.3297.6897.2998.0797.8891.9666.3882.13
Adware98.1196.6296.5297.2197.0193.0364.9583.43
Benign97.4194.9895.3895.7895.9886.6551.1872.2
Bold indicates the highest recall rate.
Table 13. F1-score comparison.
Table 13. F1-score comparison.
DatasetsCategoryModel
IZOA-LightGBMZOA-LightGBMSA-LightGBMPSO-LightGBMIZOA-XGBoostLightGBMLRMLP
CMDBenign99.8899.0399.2799.0399.2796.7767.2272.04
Adware99.7698.5698.6898.9298.297.829.8166.75
Bank Malware99.7398.6898.9499.0798.1697.8328.8736.73
SMS Malware99.8798.8798.7499.1297.9799.7247.0248.75
Riskware99.598.6198.6198.6198.3695.6249.5648.28
CCAAdware98.6896.397.1197.1195.4790.1358.4673.15
Backdoor99.7298.2998.4198.0598.2998.5371.3599.02
FileInfector98.9299.6499.5298.9299.6499.5288.19100
PUA99.7499.2299.2298.798.4597.4576.7299.05
Ransomware97.3595.6396.1594.8995.6395.6370.7580.66
Riskware99.0596.7696.7696.7696.1589.3767.1179.46
Scareware99.1399.3899.0198.7799.3899.3885.1899.26
Trojan98.7194.7294.7294.7293.7588.1464.1482.06
Trojan_Banker99.299.7399.7399.0699.699.7380.72100
Trojan_Dropper98.8898.6498.5198.2798.6498.6464.7798.14
Trojan_SMS98.4996.5797.0396.6696.5796.5766.8394.27
Trojan_Spy99.2998.5299.2698.7598.2898.2880.4695.65
AAGMGenMalware98.1796.8496.6497.2797.1391.5168.4982.12
Adware97.8296.4796.4297.1797.1691.1358.4878.4
Benign97.8595.9996.1596.6696.6188.9755.677.23
Bold indicates the highest F1-score.
Table 14. Comparison of different improvement strategies.
Table 14. Comparison of different improvement strategies.
ModelAccuracy (%)
CMDCCAAAGM
IZOA-LightGBM99.7598.8697.95
EZOA-LightGBM99.0297.9897.66
FZOA-LightGBM98.9798.1597.52
ZOA-LightGBM98.7697.7696.44
Bold indicates the highest accuracy.
Table 15. Time complexity and model size comparison.
Table 15. Time complexity and model size comparison.
DatasetsModelTraining Time (s)Detection Time (ms)Model Size (MB)
CMDIZOA-LightGBM26,426.78.5316.37
ZOA-LightGBM18,503.837.2512.69
SA-LightGBM43,180.527.1315.21
PSO-LightGBM31,810.46.5518.99
IZOA-XGBoost33,885.8813.5628.43
LightGBM3.26.891.71
LR2.677.250.26
MLP6.4215.779.03
CCAIZOA-LightGBM30,132.515.9822.23
ZOA-LightGBM25,536.287.4121.58
SA-LightGBM49,850.66.2223.66
PSO-LightGBM40,734.697.4825.87
IZOA-XGBoost30,854.3210.5637.69
LightGBM10.55.454.16
LR3.6230.37
MLP52.526.9911.53
AAGMIZOA-LightGBM28,265.454.5620.56
ZOA-LightGBM26,894.793.5521.05
SA-LightGBM50,897.626.5922.14
PSO-LightGBM42,569.174.8821.89
IZOA-XGBoost33,458.5316.5938.68
LightGBM4.533.054.27
LR1.5620.42
MLP22.664.5412.88
Bold indicates the shortest time or smallest model size.
Table 16. Comparison with other works.
Table 16. Comparison with other works.
DatasetReferenceYearAnalysisNo. of ClassesMethodAccuracy (%)
CMDMahdavifar et al. [17]2020Dynamic5PLDNN97.84
Mohamed et al. [45]2021Static2NB, SVM, KNN, DT86
Ahmed et al. [54]2022Hybrid5RF, MLP97.5
Musikawan et al. [40]2022Static2DNN98.18
Dynamic2DNN93.5
Ullah et al. [46]2022Dynamic4NB, SVM, DT, LR, RF99.11
Padmavathi et al. [55]2022Dynamic5K-means, PCA88
Jundi et al. [47]2023Hybrid5XGBoost, GE98
Tang et al. [49]2024Hybrid2HBI, DNN-AM98.67
Proposed2024Dynamic5IZOA-LightGBM99.75
CCARahali et al. [37]2020Dynamic12Semi-Supervised Deep Learning93.36
Batouche et al. [56]2021Static14RF89
Musikawan et al. [40]2022Static2DNN97.72
Dynamic14DNN78.82
Al-Andoli et al. [50]2022Static12PDL-FEMC97.6
Xie et al. [48]2023Static15MLD-Model83.17
Islam et al. [57]2023Dynamic12Ensemble ML95
Li et al. [51]2024Static12SynDroid-RF94.31
Huang et al. [39]2024Hybrid15RF88.2
Proposed2024Dynamic12IZOA-LightGBM98.86
AAGMBovenzi et al. [58]2022Dynamic3RF97
Alani et al. [52]2022Dynamic2AdStop97.08
Ullah et al. [53]2024Dynamic3FL93.85
Proposed2024Dynamic3IZOA-LightGBM97.79
Bold indicates the highest accuracy.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, S.; Li, H.; Fu, X.; Han, D.; He, X. Novel Multi-Classification Dynamic Detection Model for Android Malware Based on Improved Zebra Optimization Algorithm and LightGBM. Sensors 2024, 24, 5975. https://doi.org/10.3390/s24185975

AMA Style

Zhou S, Li H, Fu X, Han D, He X. Novel Multi-Classification Dynamic Detection Model for Android Malware Based on Improved Zebra Optimization Algorithm and LightGBM. Sensors. 2024; 24(18):5975. https://doi.org/10.3390/s24185975

Chicago/Turabian Style

Zhou, Shuncheng, Honghui Li, Xueliang Fu, Daoqi Han, and Xin He. 2024. "Novel Multi-Classification Dynamic Detection Model for Android Malware Based on Improved Zebra Optimization Algorithm and LightGBM" Sensors 24, no. 18: 5975. https://doi.org/10.3390/s24185975

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop