Next Article in Journal
Review of IoT Sensor Systems Used for Monitoring the Road Infrastructure
Next Article in Special Issue
Multi-Stage Learning Framework Using Convolutional Neural Network and Decision Tree-Based Classification for Detection of DDoS Pandemic Attacks in SDN-Based SCADA Systems
Previous Article in Journal
Road Feature Detection for Advance Driver Assistance System Using Deep Learning
Previous Article in Special Issue
An Insight into the Machine-Learning-Based Fileless Malware Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

E2E-RDS: Efficient End-to-End Ransomware Detection System Based on Static-Based ML and Vision-Based DL Approaches

by
Iman Almomani
1,2,*,†,
Aala Alkhayer
2,† and
Walid El-Shafai
2,3,†
1
Computer Science Department, King Abdullah II School for Information Technology, The University of Jordan, Amman 11942, Jordan
2
Security Engineering Laboratory, Computer Science Department, Prince Sultan University, Riyadh 11586, Saudi Arabia
3
Department of Electronics and Electrical Communications Engineering, Faculty of Electronic Engineering, Menoufia University, Menouf 32952, Egypt
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(9), 4467; https://doi.org/10.3390/s23094467
Submission received: 4 March 2023 / Revised: 18 April 2023 / Accepted: 1 May 2023 / Published: 4 May 2023
(This article belongs to the Special Issue Intelligent Solutions for Cybersecurity)

Abstract

:
Nowadays, ransomware is considered one of the most critical cyber-malware categories. In recent years various malware detection and classification approaches have been proposed to analyze and explore malicious software precisely. Malware originators implement innovative techniques to bypass existing security solutions. This paper introduces an efficient End-to-End Ransomware Detection System (E2E-RDS) that comprehensively utilizes existing Ransomware Detection (RD) approaches. E2E-RDS considers reverse engineering the ransomware code to parse its features and extract the important ones for prediction purposes, as in the case of static-based RD. Moreover, E2E-RDS can keep the ransomware in its executable format, convert it to an image, and then analyze it, as in the case of vision-based RD. In the static-based RD approach, the extracted features are forwarded to eight various ML models to test their detection efficiency. In the vision-based RD approach, the binary executable files of the benign and ransomware apps are converted into a 2D visual (color and gray) images. Then, these images are forwarded to 19 different Convolutional Neural Network (CNN) models while exploiting the substantial advantages of Fine-Tuning (FT) and Transfer Learning (TL) processes to differentiate ransomware apps from benign apps. The main benefit of the vision-based approach is that it can efficiently detect and identify ransomware with high accuracy without using data augmentation or complicated feature extraction processes. Extensive simulations and performance analyses using various evaluation metrics for the proposed E2E-RDS were investigated using a newly collected balanced dataset that composes 500 benign and 500 ransomware apps. The obtained outcomes demonstrate that the static-based RD approach using the AB (Ada Boost) model achieved high classification accuracy compared to other examined ML models, which reached 97%. While the vision-based RD approach achieved high classification accuracy, reaching 99.5% for the FT ResNet50 CNN model. It is declared that the vision-based RD approach is more cost-effective, powerful, and efficient in detecting ransomware than the static-based RD approach by avoiding feature engineering processes. Overall, E2E-RDS is a versatile solution for end-to-end ransomware detection that has proven its high efficiency from computational and accuracy perspectives, making it a promising solution for real-time ransomware detection in various systems.

1. Introduction

One of the most prevalent types of malicious software is ransomware [1,2]. Ransomware hijacks the user’s device by locking the device or encrypting the user’s data, requesting monetary payment in exchange for the locked resources, often causing irreversible information losses and enduring high economic costs. According to [3], ransomware threat attacks surpassed all other types of cybersecurity threats with around 304.7 million attacks in the first half of 2022. Therefore, there is an urgent need to develop ransomware detection systems to face the growing scourge of ransomware applications. Ransomware detection systems identify whether an application is ransomware or benign. Presently, most anti-virus software implements the signature-based approach [4]. Even though signature-based schemes consume fewer resources and less time, they are vulnerable to zero-day attacks. Additionally, the signature-based systems lack the ability to identify ransomware applications that utilize obfuscation and polymorphism techniques. Furthermore, the presence of ransomware development tool kits such as RaaS (Ransomware as a Service) contributes to the growing number of ransomware attacks [5].
To address the aforementioned issues, Machine Learning (ML) techniques are utilized in implementing ransomware analysis schemes [1,2,6,7,8]. In ransomware analysis, the functionality of a certain application is determined to classify ransomware and benign apps. The process of ransomware classification generally falls into three categories: static analysis, dynamic analysis, and vision-based analysis. In static analysis, the ransomware APK is reverse-engineered in order to retrieve the source code, strings, and resources of the application [9]. This approach lacks the ability to monitor the suspicious behavior of the ransomware since it does not involve an APK execution process. Consequently, the source code of the ransomware might bypass the static analysis even though it contains harmful methods that are invoked only during the run time process. Dynamic analysis detects ransomware applications by executing their APKs in order to observe their run-time behaviors. As a result, several features can be investigated through the run-time analysis, such as network traffic, CPU and memory usage, and system calls [10]. Dynamic analysis is typically performed in a controlled virtual environment, such as emulators, to avoid harming real devices, which might affect ransomware behavior.
In the vision-based RD analysis, the Android apps are converted to visual images before in-depth training and testing mechanisms using the developed deep learning (DL) classifiers [11,12,13,14,15]. The main improvements of vision-based analysis compared to other static-based or dynamic-based RD detectors are that they lessen the computational cost and avert reverse engineering steps needed in static-based RD analysis [16]. Moreover, they do not require to operate particular and isolated running ecosystems to check the behavior of the ransomware apps as required in dynamic-based RD analysis [17].
The research problem addressed in this paper is the need for an effective ransomware detection system to identify and prevent attacks on Android mobile systems. So, we proposed an end-to-end ransomware detection system that used a combination of machine learning and deep learning approaches based on static-based and vision-based analyses to detect ransomware attacks. Thus, this research aims to develop a system that can detect ransomware attacks in real-time on Android operating systems.
Consequently, the research problem behind this work is the growing threat of ransomware attacks on mobile systems and the need for an effective and efficient ransomware detection system. This motivated us to propose a system combining static-based machine learning and vision-based deep learning approaches to detect ransomware attacks. This is because most existing related works apply only one of these approaches in their solutions. They did not try to benefit from the advantages of both approaches within the same context. As a result, we developed a system that can identify ransomware attacks accurately and quickly, allowing for timely responses and reducing the impact of such attacks on organizations and individuals. Consequently, the static and dynamic analyses are built on feature engineering in which a database of malware features is deployed to implement the machine learning classification scheme. The rapid development of malware with constant changes raises the challenge of manually updating the malware feature database. Therefore, to reduce the cost of malware feature engineering, visual-based approaches are utilized to develop malware detection systems. In visual-based systems such as convolutional neural networks (CNN), the feature engineering process is embedded within the model construction phase, as shown in Figure 1.
The conventional detection techniques might not be efficient enough to detect new malware families, including ransomware, without causing high computational processing and memory overheads. Additionally, the traditional malware detection techniques disclose detection performance degradation due to a limited number of malware samples. Consequently, the key contributions of this paper are summarized as follows:
  • Proposing an efficient, comprehensive end-to-end RD system (E2E-RDS) that composes different static-based ML and vision-based DL classifiers.
  • Developing eight different ML models and 19 different DL models and exhaustively checking their ransomware detection capabilities.
  • Examining the proposed E2E-RDS classification performance using static and vision-based RD approaches.
  • Testing the proposed vision-based RD approach’s performance using different color and grayscale visual data of the ransomware and begin apps.
  • Assessing the proposed E2E-RDS’s performance using various detection and security evaluation parameters.
The rest of this paper proceeds as follows. Section 2 summarizes the most recently published and relevant static-based, dynamic-based, and vision-based RD systems in the literature studies. Section 3 explains the proposed end-to-end RD system. Section 4 offers the obtained simulation results and comparisons. Section 5 gives the concluding remarks and some future suggestions.

2. Literature Survey

The advent of Android ransomware applications that threaten security measures has motivated several analysis approaches for ransomware classification [18]. This section presents an overview of ransomware detection techniques used in related studies. Table 1 presents and analyzes the recent and relevant RD systems published in the literature studies. So, Table 1 compares selected papers published in 2020 and after. This range witnessed the flourishing of visual-based malware detection. Therefore, this literature survey mainly discusses three approaches recently utilized in Android ransomware detection, which are the static-based, dynamic-based, and visual-based approaches. Table 1 summarizes the comparative analysis among these approaches in terms of their aim, the dataset used, the approach followed, the ML/DL models utilized, the system input, features considered, and their solution performance concerning accuracy.

2.1. Static-Based Analysis

Static analysis efficiently implements ransomware detection systems without executing the malware code. Alkhayer et al. proposed a static analysis framework in which the decompiled APK is parsed in order to build a data set of the applications features [9,19]. Initially, the decompilation process of the APKs retrieves the Manifest file and the .smali files. After that, these files are scanned via a parsing tool, ASParse, resulting in the creation of a feature dataset. In [20], a permission-based detection system was developed to determine Android permissions that contribute to identifying ransomware with high accuracy. Zhang et al. constructed a static framework that implements a self-attention detection approach based on opcode sequence in order to classify ransomware [21].
Almomani et al. investigated the most used features of the latest version of Android (version 11) by ransomware [1]. The parsed features were utilized in building a ransomware detection system by implementing Random Forest (RF), Naive Bayes (NB) Sequential Minimal Optimization (SMO), and Decision Tree (DT). In [2], the authors simulated the real-word by implementing a hybrid evolutionary technique utilizing a highly imbalanced dataset. The dataset considered by this research consists of 10,153 Android apps, where only about 5% of the collected applications were ransomware.
In [22], an improved framework named for zero-day ransomware detection was proposed. The framework employed deep learning-based unsupervised feature extraction to extract features from the input data and a cost-sensitive Pareto Ensemble classifier to classify the data into either ransomware or benign. The cost-sensitive approach assigned different costs to misclassification errors to address the issue of imbalanced data distribution. The proposed framework showed better performance in detecting zero-day ransomware compared to traditional classifiers, highlighting the effectiveness of the approach.
Table 1. Summary of existing malware/ransomware detection schemes.
Table 1. Summary of existing malware/ransomware detection schemes.
WorkAimDatasetApproachModelsSystem InputFeaturesAccuracy
[10]To extract system calls by developing a dynamic analysis. The extracted class were fed to different ML algorithms.400 B, 400 RDynamicRandom Forest, J48, Naïve BayesAPKsystem calls98.31%
[9]To propose a static analysis in which the decompiled APK is parsed in order to build a data set of the application’s features.-Static-Decompiled APKPermissions, API calls-
[20]To determine Android permissions that contribute to identifying ransomware with high accuracy.500 B, 500 MStaticRF, J48, SMO, Naive BayesDecompiled APKPermissions96.9%
[23]To implement malware classification on a multi-class level by combining static feature extraction and deep learning.200 KB, 200 KMVisualScratch modelDecompiled APKset of stactic features93.36%
[24]To dynamically analyze an application by intercepting its network traffic.62 B, 130 MDynamicstring comparisonAPKSystem calls92%
[25]To develop a ransomware detection system by observing the behavior of system calls.502 B, 500 MDynamic-APKSystem calls98.6%
[21]To develop a self-attention detection system based on opcode sequence.100 B, 1787 MStatic, DLN-gram opcodesAPKOpcode sequence89.5%
[26]To develop a hybrid-based ransomware detection system that combines both static and dynamic analysis.500 B, 500 MHybrid (static, dynamic)-APKAPI calls-
[1]To investigate the most used features by ransomware and utilize them in building a ransomware detection system.501 B, 500 MStaticRF, DT, SMO, NBAPKPermissions, API calls98.3%
[2]To classify ransomware utilizing an evolutionary approach by deploying the SVM algorithm.9653 B, 500 MStaticparticle swarm optimization algorithm, SVM, SMOTEAPKPermissions, API calls-
[27]To improve the malware detection accuracy by deploying fine-tuned CNN models.9341 MCNNnine different modelsGray-scale image of APK-99.97%
[28]To analyze the local binary pattern by detecting the irregularity of the image texture enabling immediate detection before execution.500 B, 305 MCNNlocal binary pattern (LBP)-based technique, SVMGray-scale image of APK-87.9%
[29]To effectively classify packed and unpacked malicious software by deploying an ensemble-based CNN scheme.9339 MCNNVGG16, ResNet-50Gray-scale image of APK-99%
[30]To investigate the effectiveness of the visual-based approach by employing twelve different CNN models on a large dataset.12,971 B 20,199 MCNNVGG3, ResNet-50Gray-scale, colored, Markov, Gabor Markov images-99.97%
[31]To construct a Markov image that contains the malware statistics by deploying bytes transfer matrices.4020 MCNNScratch modelMarkov image of .dex file-97.364%
In [32], the authors proposed a new approach named spline interpolation-envisioned neural network-based for detecting ransomware. This approach combined spline interpolation and neural network techniques to improve accuracy and security. The simulation results proved the security of the proposed approach against various possible attacks. In [33], a feature selection-based ransomware detection model is presented. This model utilized information gain and a genetic algorithm to generate feature sets that can be used with any machine learning classifier. Different evaluation experiments were conducted on a real-world dataset, demonstrating the feasibility of the proposed model for ransomware detection.
In [34], the authors proposed a new optimized ML method for ransomware detection in an Internet of Things (IoT) environment. This method used the dwarf mongoose optimization algorithm to optimize the selection of parameters in an extreme ML classifier. The proposed method was evaluated on a benchmark dataset and outperformed other state-of-the-art methods in terms of accuracy and speed.

2.2. Dynamic-Based Analysis

An example of the dynamic analysis scheme is ShadowDroid which is a system that dynamically analyzes an application by intercepting its network traffic [24]. To collect the needed data, ShadowDroid set up a VPN on the targeted device and performed a string-matching algorithm to detect private information. The proposed system aims to enable users to take the correct response strategy. In [26], a hybrid approach was followed in which both static and dynamic analyses were implemented.
Abdullah et al. developed an Android ransomware classification system using dynamic-based analysis [10]. The proposed system executed the ransomware application on a virtual environment to extract the system calls. Subsequently, the extracted system calls were fed into three different machine learning algorithms, Random Forest, Naïve Bayes, and J48. Random Forest exceeded the other algorithms by achieving an accuracy of 98.31%. Furthermore, Wen et al. built a ransomware detection system based on observing system calls generated by the ransomware behaviors [25]. The proposed system focused only on the encryption-type ransomware family.
In [35], the authors presented a comprehensive analysis of techniques to evade ransomware detection techniques that use behavioral classifiers. These techniques can evade behavioral features commonly used by classifiers to detect malware, including features that are difficult to disguise and are intrinsically related to the behavior of malware processes. The authors evaluated the effectiveness of these techniques against state-of-the-art ransomware detection methods and showed that the proposed evasion techniques could effectively evade behavioral classifiers.
In [36], the authors proposed a dynamic feature dataset for detecting ransomware using machine learning algorithms. The dataset included features related to classification, encryptor, locker, and other relevant factors. The authors evaluated the dataset using three machine learning algorithms: gradient-boosted regression trees, random forest, and neural networks. They found that all three achieved an average accuracy of over 0.98 using a 10-fold cross-evaluation.

2.3. Visual-Based Analysis

Recently, deep learning has arisen as a distinguished malware classification approach. However, there is a shortage of research on detecting ransomware deploying a visual-based approach [28,37,38]. Hence, general malware texture-based classification systems were reviewed. Sharma et al. analyzed local binary patterns by detecting the irregularity of the image texture, enabling immediate detection before execution [28]. Consequently, the proposed system exposes any malicious operations injected into the ransomware byte code.
In some research works, pre-trained models can be utilized by implementing a fine-tuning phase in which the knowledge of these models is transferred. For example, El-Shafai et al. deployed fine-tuned CNN models to improve the malware detection accuracy without constructing training models [27,39]. The visual-based approach might be combined with the static approach in order to implement a multi-class classification of malware applications. In [23], the authors have proposed DIDroid, which converts the static features into 2D images aiming to perform malware characterization on a multi-class level. At first, the APKs were decompiled to extract the required features. Hence, the extracted features were converted into 2D images and fed to the deep learning model.
To effectively classify packed and unpacked malicious software, the authors of [29] deployed VGG16 and ResNet-50 CNN models utilizing the Malimg dataset to develop the proposed detection system. Pinhero et al. investigated the effectiveness of the vision-based approach by employing twelve different CNN architectures by modifying the VGG3 and ResNet-50 models [30]. A further step towards utilizing the vision-based approach in implementing malware classification schemes on the byte level was proposed by [31]. In [31], the authors constructed a Markov image containing malware statistics by deploying bytes transfer matrices.
In [40], a method for evading ransomware detection technologies that use entropy measurements to identify encrypted data was proposed. This method involved the application of various encoding algorithms, including base64 and different file formats, to alter the randomness of the data and make it appear less random, thereby bypassing the entropy-based detection methods used by ransomware detection technologies. While the proposed method was found to be effective against several ransomware detection tools, the authors caution that it may not be effective against all detection technologies, and further research is needed to develop more robust detection techniques. In [41], the authors presented a new method for detecting ransomware using graph embedding to represent the portable executable header. The proposed approach overcame the limitations of traditional signature-based and machine learning-based methods that require large amounts of labeled data. The authors evaluated their method on a dataset of over 2000 ransomware and non-ransomware samples and found that it achieved high accuracy and outperformed existing methods.
As can be observed from Table 1, no existing study has considered both static/dynamic and vision-based approaches within the same context to examine their performance in detecting malware in general and ransomware in specific. Consequently, this has motivated this research to conduct a deep study on the impact of applying both approaches while considering the same malware datasets and implementation environments.

3. Proposed End-to-End Ransomware Detection System

This section discusses the details of the proposed end-to-end ransomware detection system. End-to-end means full utilization of different types of analysis methods, whether the ones that depend on reverse engineering where the malware code is recovered and parsed and several features are gathered, such as the case in the static-based analysis or by taking the malware executable code as is (PE: Portable Executable), converting it to an image, and then applying analysis as in the case of vision-based systems. The aim is to investigate the advantage of each one of them to ensure efficient detection of ransomware apps using the same inputs and environments.
Figure 2 shows the high-level diagram of the E2D-RDS system that composes two proposed approaches to efficiently detect ransomware apps. The first static-based RD approach has utilized ML models, as will be discussed in Section 3.1. In contrast, the second vision-based RD approach has used different CNN models by exploiting the substantial advantages of FT and TL mechanisms, as will be clarified in Section 3.2.
Here are the main steps of the proposed E2E-RDS system that are presented in Figure 2:
  • Load dataset containing ransomware and non-ransomware files.
  • Preprocess the dataset by extracting features.
  • Split extracted features dataset into training and testing sets.
  • Train static-based machine learning models on the training set.
  • Test the machine learning models on the testing set.
  • Convert the dataset samples into color and gray images.
  • Split images dataset into training and testing sets.
  • Train vision-based deep learning models on the training set.
  • Test the deep learning models on the testing set.
  • Check the output of the static-based and vision-based models.
  • If the output indicates ransomware, flag the file as malicious.
  • If the output indicates non-ransomware, flag the file as safe.
  • Evaluate the performance of the system using different relevant metrics such as accuracy, precision, recall, and other detection metrics.
  • Fine-tune the models and retrain if necessary.
  • Save the final models for use in future detections.

3.1. The Proposed Static-Based RD Approach

As indicated in Figure 2, in the static-based RD approach, the binary portable APK files are first decompiled to extract the AndroidManifest.xml binary file, which comprises the APK’s metadata and all well-defined permissions by the Android APKs. In the decompile process, we used the APKtool (https://ibotpeaches.github.io/Apktool/, accessed on 1 January 2020) to decompile the zipped Android app to the manifest and SMALI files. The SMALI file signifies a well-defined class in the original binary code of the Android APK. After that, the required features of the Android APKs are obtained by analyzing the obtained SMALI and manifest files.
In the parsing process, the obtained feature set used in the proposed static-based RD approach includes 161 permissions, 228 API packages, and 389 features. The feature occurrences have been calculated using the assembled APKs by obtaining the features from the manifest and SMALI files. Therefore, all Android APKs are scanned using separate stages during the parsing process. In the first stage, the manifest file of each Android APK is parsed to count the specified features, such as permissions. Then, in the second stage, the SMALI files of each Android APK are parsed to count the utilized API packages. At the end, the total parsed extracted features are deposited in the database.
In addition, in the proposed static-based RD approach, we apply pre-processing, preparing, and cleaning mechanisms to the extracted features to remove the zero-values or null attributes and represent the extracted features in the proper format before forwarding them to the utilized ML classifiers. This approach uses eight different ML models to train and test the extracted features from the APK files. These ML classifiers are Support Vector Machine (SVM), Random Forest (RF), Ada Boost (AB), Decision Tree (CART), K-Nearest Neighbor (KNN), Naive Bayes (NB), Linear Discriminant Analysis (LDA), and Logistic Regression (LR) [42,43,44]. Finally, the detection efficacy of the used ML classifiers is evaluated using different assessment tools such as precision, recall, F1-score, ROC curve, confusion matrix, and accuracy [45,46].

3.2. The Proposed Vision-Based RD Approach

The efficient Ransomware Detection (RD) process is an obligatory aspect of cybersecurity applications because ransomware is a highly harmful malicious software that can infect and encrypt the users’ or organizations’ files and request payment to decipher the encrypted files. Most of the RD approaches developed recently have introduced different detection mechanisms to identify ransomware after extracting its main texture features (static-based RD) or starting its execution (dynamic-based RD). However, unfortunately, the static-based RD approaches require additional processing stages to analyze and extract the ransomware features. In addition, until recently once the ransomware app is executed and begins its attack, no currently developed dynamic-based RD approaches have been suggested that can prevent and stop its harm. Therefore, instead of using conventional static- or dynamic-based RD approaches, this section introduces a vision-based RD approach by exploiting 19 different FT CNN models. These models can inexpensively identify ransomware apps with low computation requirements and high detection performance.
In the proposed vision-based RD approach, the binary executable files of the benign and ransomware apps are converted into 2D visual (color or grayscale) images. Then, these images have been forwarded to the employed FT CNN models for a binary classification purpose to differentiate ransomware apps from benign apps. The main benefit of this approach is that it can efficiently detect and identify ransomware without using data augmentation or complicated feature extraction processes. Therefore, the FT (Fine-Tuning) and TL (Transfer Learning) techniques were only exploited in the proposed vision-based RD approach to achieve high detection accuracy. The TL process offers effective and promising solutions through the knowledge transfer of pre-trained CNN models. While in the FT technique, the pre-trained CNN layers are progressively trained by modifying their weights and the learned hyperparameters until they achieve a remarkable classification performance.
As indicated in Figure 2, the primary strategy of the proposed vision-based RD approach involves three main stages: (1) dataset conversion and preparation stage, (2) TL, FT, and classification stage, and (3) performance assessment stage. The explanations of these stages are as follows:

3.2.1. Dataset Conversion and Preparation Stage

In this stage, the benign and ransomware apps of our newly collected balanced 500/500 (ransomware/benign) dataset are directly converted to both color and gray images without using any decryption, decompression, and disassembly methods. The primary goal of transforming the PE (portable executable) APKs to images is to acquire the main texture features of the benign and ransomware apps. In the conversion process, the benign and ransomware APKs are first transformed into 1D 8-bit binary vectors, and after that, these binary vectors are transformed into visual 2D color or gray images. The CNN models utilized in the proposed vision-based RD approach for the binary classification task exploited the resulting texture features extracted from the obtained images in detecting ransomware apps efficiently. Thus, the significant benefit of the proposed vision-based RD approach is that it does not need any feature extraction or reverse-engineering tasks to be performed.
Figure 3 demonstrates some visual color and gray images of the ransomware and benign apps after reorganizing the 1D binary vectors into 2D graphical arrangements. It is noticed from Figure 3 that the texture features and stripes of the obtained ransomware images are entirely different from the texture features and stripes of the benign images, either in the color or gray images. Additionally, as observed, each converted image has a different width than another image because each benign or ransomware app has a different size. So, the obtained image width is based on the size of the benign and ransomware APKs. Table 2 shows the typical different image widths of benign and ransomware APKs based on their sizes. Consequently, such remarks have inspired us to optimize and exploit the common pre-trained CNN architectures applied for image recognition and classification tasks into vision-based RD tasks.
Before forwarding the accumulated visual images of the ransomware and benign apps to the suggested fine-tuned CNN models for detection and classification purposes, these images are divided into 80% to train the layers of CNN models and 20% for testing the CNN models. These testing and training percentages of benign and ransomware images are selected randomly for the accumulated visual image datasets (color and grayscale). Additionally, as a preparation step, the obtained visual images of the benign and ransomware apps must be resized before performing the training and classification processes using a specified CNN model. This is because each one of the suggested fine-tuned CNN models has its standard size for the input visual images, as demonstrated in Table 3.

3.2.2. TL, FT, and Classification Stage

As shown in Figure 2, the proposed vision-based RD approach comprises 19 various CNN models, which are ResNet50, AlexNet, InceptionV3, ResNet101, GoogleNet, VGG16, DarkNet53, Xception, InceptionResNetV2, MobileNetV2, NasNetMobile, DarkNet19, ResNet18, DenseNet201, NasNetLarge, Places365-GoogleNet, ShuffleNet, SqueezeNet, and VGG19 [47,48,49]. These models do not employ reverse engineering to extract the main features of the visual images used in the binary classification process of detecting and recognizing ransomware apps. Thus, in the proposed vision-based RD approach, we exploit the transfer learned features of the fine-tuned and optimized CNN versions of these pre-trained CNN models to detect ransomware attacks without the need to design deep CNN models learned from scratch.
The CNN models used in the proposed vision-based RD approach can be classified into two categories: (1) single-path designs and (2) multi-path designs. In the first category of single-path designs, the CNN layers are arranged in a series path with a sequential structure, and they have single input and output layers. In the second category of multi-path designs, there are multiple parallel paths for the composed CNN layers in the utilized CNN model, and thus they have multiple input and output layers. Examples of single-path CNN designs are AlexNet, VGG16, DarkNet19, and VGG19 models, wheres examples for the multi-path CNN designs are ResNet50, InceptionV3, ResNet101, GoogleNet, DarkNet53, Xception, InceptionResNetV2, MobileNetV2, NasNetMobile, ResNet18, DenseNet201, NasNetLarge, Places365-GoogleNet, ShuffleNet, and SqueezeNet models [47,48,49]. From a complexity perspective, the multi-path designs-based CNN models are more complex than the single-path-design-based CNN models. They have an additional advantage in attaining high detection accuracy and a low misclassification rate. This is due to the extensive extracted features from the input visual ransomware and benign images by different composed CNN layers during the training process.
The TL is considered a supervised learning process that trains a CNN model on one classification problem with a specific image dataset. After that, it utilizes the learned features in another and different classification challenge on a different image dataset. The significant profit of using the TL process is to decrease the examined CNN model’s training time and achieve low generalization and classification errors, thus avoiding the overfitting occurrence. In addition, the importance of employing the TL process is highlighted when the dataset used in the classification task has few samples of visual images, as in our proposed vision-based RD approach (500 ransomware samples and 500 benign samples).
Amongst the nineteen different tested CNN structures utilized in the proposed vision-based RD approach, the optimized version of the ResNet50 CNN structure attains the best detection efficiency and classification accuracy. Therefore, in our work, in-detail discussions and explanations of this CNN model (FT ResNet50) are presented. The general details and basic descriptions of the other 18 different CNN structures (AlexNet, InceptionV3, ResNet101, GoogleNet, VGG16, DarkNet53, Xception, InceptionResNetV2, MobileNetV2, NasNetMobile, DarkNet19, ResNet18, DenseNet201, NasNetLarge, Places365-GoogleNet, ShuffleNet, SqueezeNet, and VGG19) used in the proposed vision-based RD approach could be explored in [50,51,52,53,54,55,56].
The structural design of the employed FT ResNet50 CNN model employed in the proposed RD approach is depicted in Figure 4. The pre-trained version of this ResNet50 CNN model is trained on more than ten million natural images with more than 1000 various classes composed in the common ImageNet dataset [57]. Thus, in the proposed vision-based RD approach, we exploited the resulting learned features and performed fine-tuning for the convolutional layers’ hyperparameters and weights to detect ransomware attacks efficiently and rapidly.
The utilized residual network (ResNet50) has a simpler multi-path design structure than other multi-path CNN structures. It consists of four different sequential stages of convolutional (Conv) layers with different sizes and numbers of integrated filters, as shown in Figure 4. These Conv layers extract the main features from the input visual ransomware and benign images. Each of these Conv layers in the ResNet50 structure has a batch normalization layer and a Rectified Linear Unit (ReLU) activation function that is not indicated in Figure 4 for presentation simplicity. So, there is a hidden ReLU function after each hidden batch normalization layer for each Conv layer in Figure 4. The ReLU functions are used to activate the nonlinear batch normalization layers for quickly accomplishing convergence performance and thus accelerating the training mechanism. Therefore, the ResNet50 model is faster and more accurate than other multi-path CNN structures. In addition, it can extract more features and characteristics from the input visual images of the ransomware and benign apps.
Furthermore, the ResNet50 model includes a max-pooling layer with a kernel size of 3 × 3 at the input level, and it incorporates an average pooling layer with a kernel size of 7 × 7 at the output level. The max-pooling layer determines the maximum value within the patches of the extracted feature maps. In contrast, the average pooling layer is employed to determine the average value within the patches of the extracted feature maps. Finally, the extracted features resulting from the Conv layers are forwarded to the classification layers (fully connected and softmax layers) to differentiate the ransomware samples from the benign samples. In the FT ResNet50 model, the last fully connected layer is modified to have two outputs (ransomware and benign) instead of the 1000 outputs as in the original pre-trained ResNet50 CNN model.
The main advantages of the employed FT ResNet50 model are (1) the utilization of the batch normalization layer that optimizes the parameters of the input layer to improve the CNN model performance, and thus the covariate shift is alleviated, (2) the utilization of the identity connections that protect the CNN model structure from diminishing gradient problems, (3) the utilization of residual bottleneck block designs that enhance the CNN model performance, and (4) the utilization of different sizes of kernels within the Conv layers, and thus the achievement of deep discovering and learning of the foremost characteristics and texture features from the input visual images.
As discussed in the proposed vision-based RD approach, the fine-tuning for the hyperparameters of the layers included in the employed CNN models is carried out besides exploiting TL benefits. The fine-tuning method is preferable during the training of CNN models compared to other tunning methods (e.g., shallow tuning and deep tuning, because the FT achieves higher detection performance and lower computations than shallow and deep tuning approaches). Thus, this motivates us to use the FT process in the proposed vision-based RD approach to optimize the weights and parameters of the CNN layers until achieving optimum classification results for efficiently detecting the ransomware attacks in android mobile operating systems. The optimum fine-tuning and training parameters utilized in the second proposed approach are summarized in Table 4. These optimization parameters are thoughtfully selected after performing different simulation experiments on the examined CNN models using the created visual android dataset until optimizing the testing and training operations of the employed CNN models to avoid overfitting problems.

3.2.3. Performance Assessment Stage

Various detection assessment metrics have been exhaustively introduced in the literature to offer comprehensive evaluations of detection and classification algorithms [45,46]. In our performance analysis of the employed CNN models, we used various detection metrics instead of only evaluating the detection accuracy, since the utilization of only detection accuracy as an assessment metric of the classification performance is not precise from conceptual and practical perspectives. So, the detection accuracy combines the result in the competence of the CNN model to appropriately predict both N (negative) and P (positive) cases, and it is not potential to assess if the CNN model is better in accurately expecting N or P. Thus, a high accuracy value can be achieved by the excellent ability of the CNN model to predict only one of the possible classes accurately.
Therefore, when the performance accomplishment of a CNN model is assessed, other evaluation metrics of detection quality are also necessary to be estimated. Moreover, those evaluation metrics are very beneficial for comparing various CNN-based classification models because different CNN models can expect very distinct classes but with identical classification accuracy. Thus, different detection analysis metrics are exploited in this stage to explore the performance assessment of the developed nineteen CNN models for our investigated binary classification challenge of detecting ransomware attacks.
Our detection analysis of the binary classification challenge has been performed based on obtaining the results of the confusion matrix, loss and accuracy curves, validation accuracy, true-positive rate (TPR) (recall) (sensitivity), predictive value (PPV) (precision), negative predictive value (NPV), true-negative rate (TNR) (specificity), F1-Score, area under the receiver operating characteristic (AROC) curve, AROC score, false-negative rate (FNR), false-positive rate (FPR), false-omission rate (FOR), false-discovery rate (FDR), and misclassification rate.
In any binary classification challenge, if N (negative) and P (positive) are the probable classes/labels for each examination, a detection CNN model can generate only four scores, as shown in Figure 5. These scores are (1) TP (true positive): this score occurs if the CNN model predicts P, and it is likewise the true response, (2) FP (false positive): this score occurs if the CNN model predicts P, but the true response is N, (3) TN (true negative): this score occurs if the CNN model predicts N, and it is likewise the true response, and (4) FN (false negative): this score occurs if the CNN model predicts N, but the true response is P. These four scores can be utilized to estimate the numerical values of the examined detection assessment metrics; their mathematical formulas are expressed as follows:
Validation accuracy = T N + T P F P + T P + F N + T N
TPR ( sensitivity ) = recall = T P F N + T P
PPV ( precision ) = T P F P + T P
NPV = T N F N + T N
TNR ( specificity ) = T N F P + T N
F 1 - Score = 2 TP 2 TP + FN + FP
false - negative rate ( FNR ) = F N T P + F N
false - positive rate ( FPR ) = F P T N + F P
false - omission rate ( FOR ) = F N T N + F N
false - discovery rate ( FDR ) = F P T P + F P
misclassification rate = F N + F P F P + T P + F N + T N
The loss curve is a tracing representation that demonstrates the estimated loss ratios for all examined training epochs, while the accuracy curve is a tracing representation that indicates the estimated accuracy ratios for all examined training epochs. Finally, the AROC curve represents the trade-off between the FPR (1-specificity) and TPR, while the AROC score is the estimated area under the AROC curve.

4. Results and Comparisons

This section introduces the analysis and discussions of the proposed E2E-RDS, including static-based and vision-based RD approaches using our collected and balanced benign and ransomware APKs. The dataset used in this work is publicly available on the Security Engineering Lab (SEL) website (https://sel.psu.edu.sa/Research/datasets/2020_RansIm-DS.php, accessed on 1 January 2020). The collection process of this dataset started by assembling the Android APKs (ransomware and benign). Both ransomware and benign samples were downloaded as Android Packages (APK). Benign APKs were downloaded from the Google play store. While ransomware APKs were downloaded from several famous market repositories, including HelDroid, VirusTotal, RansomProper, and Koodous. After that, the duplicated Android apps were removed.

4.1. Performance Analysis and Discussion of the Proposed Static-Based RD Approach

This section discusses the performance analysis of the proposed static-based RD approach. The evaluation of the utilized eight different ML classifiers is performed on the extracted features of the balanced ransomware and benign images (500 ransomware and 500 benign). These features are divided into 80% to train the layers of the examined ML classifiers and 20% for testing the utilized ML classifiers. These testing and training percentages of benign and ransomware features are selected randomly from the obtained static features dataset.
For simplicity in the presentation of the results, the obtained confusion matrix and ROC curve of only the best proficient ML classifier used in the proposed static-based RD approach are presented. At the same time, the summarized outcomes of the average values of all quantitive evaluation metrics (accuracy, recall, precision, and F1-Score) are offered for all examined ML classifiers.
Figure 6 declares the obtained confusion matric and ROC curve for the best accomplished AB classification model used in the proposed static-based RD approach. The acquired confusion matrix and ROC curve reveal that the AB ML-based RD classifier can correctly detect almost all ransomware APKs and differentiate them from benign APKs using the static input features.
The performance analysis results of the average values of different classification assessment metrics for the examined ML classifiers are given in Table 5. The tested ML models utilized in the proposed static-based RD approach are observed to achieve reasonable and acceptable detection outcomes of high detection accuracy, F1-Score, recall, and precision values. The tested ML models utilized in the proposed static-based RD approach are observed to achieve reasonable and acceptable detection outcomes of high detection accuracy, F1-Score, recall, and precision values. The best-accomplished ML classifier that achieves high detection accuracy was the AB model, while the LR model achieved the lowest detection performance compared to other examined ML models.

4.2. Performance Analysis and Discussion of the Proposed Vision-Based RD Approach

In this section, the performance analysis of the proposed vision-based RD approach is discussed. The training and validation experiments were carried out using MATLAB 2020b on a laptop machine with an Intel Core i7-4500 processor with 8 gigabyte memory without a GPU accelerator. The evaluation of the 19 different employed CNN models is performed on the created balanced visual ransomware and benign images (500 ransomware images and 500 benign images). These images were divided into 80% to train the layers of the examined CNN models and 20% for testing the utilized CNN models. These testing and training percentages of benign and ransomware images were selected randomly from the created visual color or gray dataset.
Therefore, the detection efficiency and classification performance of the proposed vision-based RD approach that composes nineteen different employed CNN models have been examined using both visual color and grayscale images generated from our collected ransomware and benign apps. This is to extensively analyze and investigate the effect of using two different visual image representations (color and grayscale formats) of the android ransomware and benign APKs on the detection efficiency of the proposed vision-based RD approach.
For simplicity in the presentation of the results, the detailed outcomes regarding the confusion matrix, loss and accuracy curves, and AROC curves of the best proficient fine-tuned ResNet50 model in the proposed vision-based RD approach are presented. While the summarized outcomes of the average values of all quantitive evaluation metrics (validation accuracy (Val. Acc.), recall (Rec.), precision (Prec.), NPV, specificity (Spec.), F1-Score, AROC score, FNR, FPR, FOR, FDR, and misclassification rate (Mis. Class. Rate)) that are discussed in Section 3.2.3 are offered for all developed and fine-tuned CNN models utilized in the proposed vision-based RD approach.
Figure 7 and Figure 8 declare the obtained accuracy and loss curves for the training and validation processes of the best accomplished ResNet50 classification model using the color and grayscale image datasets, respectively.
It can be realized that the testing and training accuracy curves for both color and grayscale datasets are matched to each other, and they tend to be steady after only five epochs. However, the testing loss curves were marginally higher than the training loss curves, especially for the grayscale image dataset, but the obtained validation loss has an acceptable average value of less than 0.1. Thus, it is demonstrated that the outcomes of the testing and training loss and accuracy curves for the visual color dataset are better than those for the visual grayscale dataset. Both visual color and grayscale datasets proved and validated the excellent detection performance of the examined fine-tuned ResNet50 model using different visual image representations for the ransomware and benign APKs. So, the fine-tuned ResNet50 model achieved high detection and low misclassification performance at minimal epochs (training iterations). Furthermore, analogous outstanding achievements of loss and accuracy curves are acquired and accomplished for the other investigated 18 different CNN models tested in the proposed vision-based RD approach for the color and grayscale image datasets. Consequently, the developed fine-tuned CNN models that have been used in the proposed vision-based RD approach are highly recommended to be exploited in cybersecurity applications for accurately detecting ransomware attacks in Android operating systems.
Figure 9 shows the obtained confusion matrices for the best accomplished FT ResNet50 classification model using color and grayscale image datasets. In addition, Figure 10 demonstrates the obtained AROC curves for the best accomplished ResNet50 classification model using color and grayscale image datasets. These acquired confusion matrices and AROC curves reveal that the fine-tuned ResNet50 model can correctly detect almost all ransomware APKs and differentiate them from benign APKs using color or grayscale image datasets. Furthermore, it is clear that the attained confusion matrices or AROC curves of the color image dataset are slightly better than those achieved using the grayscale image dataset. These similar observations are obtained for the other 18 different examined FT CNN models utilized in the proposed vision-based RD approach. Thus, the employed fine-tuned CNN models are highly recommended to be used efficiently for detecting ransomware attacks in Android cybersecurity applications.
The performance analysis results of the average values of different classification assessment metrics for the examined CNN models on the color and grayscale image datasets are given in Table 6 and Table 7, respectively. It is observed that the whole tested fine-tuned CNN models utilized in the proposed vision-based RD approach achieve promising outcomes of high validation accuracy, TPR, PPV, NPV, TNR, F1-Score, and AROC values, and also, all of them attain low FNR, FPR, FOR, FDR, and misclassification rate values for both color and grayscale datasets. Moreover, it is demonstrated that the assessment detection values obtained using the color image dataset are slightly higher than those obtained using the grayscale image dataset. Furthermore, despite its design simplicity, the fine-tuned ResNet50 CNN model achieved the best detection results compared to other fine-tuned CNN models for both color and grayscale images. These promising classification results for the whole tested FT CNN-based TL models have been accomplished with the aid of using transfer learned features resulting from the pre-trained CNN models in addition to exploiting the fine-tuning of their layers and hyperparameter values.
After validating the classification performance analysis and detection efficiency of the proposed fine-tuned CNN models, we have investigated the complexity analysis of the proposed vision-based RD approach concerning (1) the storage size of the utilized visual datasets, (2) experimental requirements of the employed CNN models, and (3) computation detection overhead of the FT CNN models. Table 8 presents the storage size in gigabytes of the color and grayscale visual images of the ransomware and benign samples, while Table 9 indicates the specifications of the utilized FT CNN models. Finally, Table 10 introduces the computational analysis of the utilized FT CNN models.
Table 8 shows that the storage size of the visual color samples (4.198 GB) has a minor increase in size compared to the storage size of the visual grayscale samples (4.143 GB). The last row in Table 8 indicates that there is a slight storage size increase of 0.055 GB for the visual color samples compared to the visual grayscale samples. Thus, there is an increase in percentage of 1.31% produced when using the visual color samples, which is reasonable and tolerable compared to the added benefits from using them compared to visual grayscale samples concerning the accomplished high detection efficiency and low misclassification rate for almost all FT CNN models as discussed and indicated previously in Table 6 and Table 7.
The disk size, total number of layers, number of parameters (total, trainable, and non-trainable), and reduction percentage in training parameters of the whole employed FT CNN models in the proposed vision-based RD approach are stated in Table 9. As indicated, these experimental requirements and specifications differ from one CNN model to another CNN model, depending on the model structure (single-path design or multi-path design) and the total number of included CNN layers. Furthermore, the number of training and non-training parameters of each CNN model is based on the number of frozen and unfrozen layers. In the proposed vision-based RD approach, most of the training parameters of the CNN layers of the utilized fine-tuned models are frozen. So, for example, the FT ResNet50 model has trainable parameters of 4,096 and non-trainable parameters of 25,595,904, and thus there is a high reduction percentage in training parameters reached 99.98%. This significant reduction in training parameters comes from the valuable utilization of transfer learned features resulting from the pre-trained CNN models and exploiting the considerable advantages of fine-tuning the hyperparameters, weights, and CNN layers of the pre-trained CNN-based TL models as indicated in Table 9, the same observations concerning the number of training parameters and their reduction percentage are obtained for all examined FT CNN models in the proposed vision-based RD approach.
The comprehensive computational analysis of the FT CNN models for the visual color and grayscale samples are introduced in Table 10. The total training and validation time and the average detection time to identify the ransomware or benign sample for each CNN model are estimated. As observed, the processing time is different from one CNN model to another CNN model, depending on the number of training parameters and CNN layers as clarified in Table 9. It can be concluded that the computational analysis of the whole tested FT CNN models proves that the average validation time spent to recognize ransomware samples or benign samples is acceptable for all FT CNN models, comparable to their accomplished classification accuracy.
The overall extensive validation of the proposed E2E-RDS has proven its efficiency in detecting ransomware, whether by reverse engineering the code itself and then analyzing it or by handling it in its current format and applying vision-based analysis. The intelligence is applied using ML or CNN models while considering the system’s resources and requirements. Finally, the main target is achieved in detecting ransomware with high accuracy to protect users’ machines and data.
In addition to the presented results, we also compared the proposed RDS system with other related works that have been evaluated on the same dataset. Table 11 shows the comparative analysis outcomes. Compared to related works evaluated on the same dataset, it is noticed that the proposed RDS system offers a high detection performance for both static-based ML and vision-based DL approaches. So, compared to the other systems, the achieved detection accuracy of the proposed system indicates its superiority in detecting ransomware.

5. Conclusions and Future Work

Ransomware still registers high numbers of successful attacks with a significant impact. Therefore, many researchers are still working to enhance the performance of ransomware detection, especially in the context of Android systems. However, the existing approaches usually utilize one of the analytical approaches that either reverse engineer the ransomware or deal with it in its current format. Both approaches were used in the literature but (a) separately by applying one of them only and (b) differently by considering various contexts, datasets, simulation environments, and evaluation metrics. This paper comprehensively analyses Android ransomware by introducing an efficient end-to-end ransomware detection system (E2E-RDS). The proposed E2E-RDS either reverses engineering the ransomware code, parsing it, extracting its essential features, and then building predictive systems that are based on machine learning (ML) classifiers, or keeps the ransomware in its executable format, converts it to images and then build predictive systems using vision-based CNN models.
The proposed E2E-RDS aims to assess the capabilities of applying different analysis models while considering the same context in terms of ransomware datasets, experimental environments, and evaluation metrics. The purpose is to highly and efficiently recognize the ransomware apps and prevent the users from installing them while considering the system resources and requirements. The experiments reveal that the resulting predictive models from different analysis systems succeeded in reaching 99.5% detection accuracy. Eight ML classifiers and 19 CNN models were developed and examined to reach this accuracy. Moreover, extensive experiments were conducted to examine not only the accuracy but many other metrics related to security and complexity. The target is to highlight all aspects related to malware detection in general and ransomware in specific to choose the proper model while building the ransomware detection systems.
In future work, different malware types other than ransomware could be investigated. Additionally, different ML and CNN models could also be integrated and examined within the E2E-RDS. Moreover, we intend to build an end-to-end mathematical analysis for the proposed RDS system. Furthermore, we target to test the detection performance of the proposed RDS system on real-world applications and services.

Author Contributions

Conceptualization, I.A. and W.E.-S.; methodology, I.A. and W.E.-S.; software, W.E.-S. and A.A.; validation, I.A., A.A. and W.E.-S.; formal analysis, I.A. and W.E.-S.; investigation, I.A., A.A. and W.E.-S.; resources, I.A., A.A. and W.E.-S.; data curation, I.A., A.A. and W.E.-S.; writing—original draft preparation, W.E.-S. and A.A.; writing—review and editing, I.A.; visualization, I.A., A.A. and W.E.-S.; supervision, I.A. and W.E.-S.; project administration, I.A. and W.E.-S.; funding acquisition, I.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. The APC is funded by Prince Sultan University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to acknowledge the support of Prince Sultan University for paying the Article Processing Charges (APC) of this publication.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Almomani, I.; AlKhayer, A.; Ahmed, M. An Efficient Machine Learning-based Approach for Android v. 11 Ransomware Detection. In Proceedings of the 2021 1st International Conference on Artificial Intelligence and Data Analytics (CAIDA), Riyadh, Saudi Arabia, 6–7 April 2021; pp. 240–244. [Google Scholar]
  2. Almomani, I.; Qaddoura, R.; Habib, M.; Alsoghyer, S.; Al Khayer, A.; Aljarah, I.; Faris, H. Android Ransomware Detection Based on a Hybrid Evolutionary Approach in the Context of Highly Imbalanced Data. IEEE Access 2021, 9, 57674–57691. [Google Scholar] [CrossRef]
  3. SonicWal. Sonicwall Cyber Threat Report. 2021. Available online: https://www.sonicwall.com/2021-cyber-threat-report/ (accessed on 4 August 2022).
  4. Al-Asli, M.; Ghaleb, T.A. Review of signature-based techniques in antivirus products. In Proceedings of the 2019 International Conference on Computer and Information Sciences (ICCIS), Sakaka, Saudi Arabia, 3–4 April 2019; pp. 1–6. [Google Scholar]
  5. Urooj, U.; Maarof, M.A.B.; Al-rimy, B.A.S. A proposed Adaptive Pre-Encryption Crypto-Ransomware Early Detection Model. In Proceedings of the 2021 3rd International Cyber Resilience Conference (CRC), Langkawi Island, Malaysia, 29–31 January 2021; pp. 1–6. [Google Scholar]
  6. Derhab, A.; Cheikhrouhou, O.; Allouch, A.; Koubaa, A.; Qureshi, B.; Ferrag, M.A.; Maglaras, L.; Khan, F.A. Internet of Drones Security: Taxonomies, Open Issues, and Future Directions. Veh. Commun. 2022, 39, 100552. [Google Scholar] [CrossRef]
  7. Alkhelaiwi, M.; Boulila, W.; Ahmad, J.; Koubaa, A.; Driss, M. An efficient approach based on privacy-preserving deep learning for satellite image classification. Remote Sens. 2021, 13, 2221. [Google Scholar] [CrossRef]
  8. Noor, A.; Zhao, Y.; Koubâa, A.; Wu, L.; Khan, R.; Abdalla, F.Y. Automated sheep facial expression classification using deep transfer learning. Comput. Electron. Agric. 2020, 175, 105528. [Google Scholar] [CrossRef]
  9. Al Khayer, A.; Almomani, I.; Elkawlak, K. ASAF: Android Static Analysis Framework. In Proceedings of the 2020 First International Conference of Smart Systems and Emerging Technologies (SMARTTECH), Riyadh, Saudi Arabia, 3–5 November 2020; pp. 197–202. [Google Scholar]
  10. Abdullah, Z.; Muhadi, F.W.; Saudi, M.M.; Hamid, I.R.A.; Foozy, C.F.M. Android Ransomware Detection Based on Dynamic Obtained Features. In Proceedings of the International Conference on Soft Computing and Data Mining, Langkawi, Malaysia, 22–23 January 2020; Springer: Cham, Switzerland, 2020; pp. 121–129. [Google Scholar]
  11. Kumar, S.; Janet, B.; Neelakantan, S. Identification of malware families using stacking of textural features and machine learning. Expert Syst. Appl. 2022, 208, 118073. [Google Scholar] [CrossRef]
  12. Bovenzi, G.; Cerasuolo, F.; Montieri, A.; Nascita, A.; Persico, V.; Pescapé, A. A comparison of machine and deep learning models for detection and classification of android malware traffic. In Proceedings of the 2022 IEEE Symposium on Computers and Communications (ISCC), Rhodes, Greece, 30 June–3 July 2022; Volume 22. [Google Scholar]
  13. Kumar, P.; Gupta, G.P.; Tripathi, R. Design of anomaly-based intrusion detection system using fog computing for IoT network. Autom. Control Comput. Sci. 2021, 55, 137–147. [Google Scholar] [CrossRef]
  14. Kumar, P.; Gupta, G.P.; Tripathi, R. Toward design of an intelligent cyber attack detection system using hybrid feature reduced approach for iot networks. Arab. J. Sci. Eng. 2021, 46, 3749–3778. [Google Scholar] [CrossRef]
  15. Kumar, P.; Gupta, G.P.; Tripathi, R. A distributed ensemble design based intrusion detection system using fog computing to protect the internet of things networks. J. Ambient Intell. Humaniz. Comput. 2021, 12, 9555–9572. [Google Scholar] [CrossRef]
  16. Awan, M.J.; Masood, O.A.; Mohammed, M.A.; Yasin, A.; Zain, A.M.; Damaševičius, R.; Abdulkareem, K.H. Image-Based Malware Classification Using VGG19 Network and Spatial Convolutional Attention. Electronics 2021, 10, 2444. [Google Scholar] [CrossRef]
  17. Ben Abdel Ouahab, I.; Elaachak, L.; Bouhorma, M. Classification of Malicious and Benign Binaries Using Visualization Technique and Machine Learning Algorithms. In Big Data Intelligence for Smart Applications; Springer: Berlin/Heidelberg, Germany, 2022; pp. 297–315. [Google Scholar]
  18. Sharma, S.; Kumar, R.; Rama Krishna, C. A survey on analysis and detection of Android ransomware. Concurr. Comput. Pract. Exp. 2021, 33, e6272. [Google Scholar] [CrossRef]
  19. Raji, I.D.; Bello-Salau, H.; Umoh, I.J.; Onumanyi, A.J.; Adegboye, M.A.; Salawudeen, A.T. Simple deterministic selection-based genetic algorithm for hyperparameter tuning of machine learning models. Appl. Sci. 2022, 12, 1186. [Google Scholar] [CrossRef]
  20. Alsoghyer, S.; Almomani, I. On the effectiveness of application permissions for Android ransomware detection. In Proceedings of the 2020 6th Conference on Data Science and Machine Learning Applications (CDMA), Riyadh, Saudi Arabia, 4–5 March 2020; pp. 94–99. [Google Scholar]
  21. Zhang, B.; Xiao, W.; Xiao, X.; Sangaiah, A.K.; Zhang, W.; Zhang, J. Ransomware classification using patch-based CNN and self-attention network on embedded N-grams of opcodes. Future Gener. Comput. Syst. 2020, 110, 708–720. [Google Scholar] [CrossRef]
  22. Zahoora, U.; Khan, A.; Rajarajan, M.; Khan, S.H.; Asam, M.; Jamal, T. Ransomware detection using deep learning based unsupervised feature extraction and a cost sensitive Pareto Ensemble classifier. Sci. Rep. 2022, 12, 15647. [Google Scholar] [CrossRef] [PubMed]
  23. Rahali, A.; Lashkari, A.H.; Kaur, G.; Taheri, L.; Gagnon, F.; Massicotte, F. DIDroid: Android Malware Classification and Characterization Using Deep Image Learning. In Proceedings of the 2020 the 10th International Conference on Communication and Network Security, Tokyo, Japan, 27–29 November 2020; pp. 70–82. [Google Scholar]
  24. Hu, C.C.; Jeng, T.H.; Chen, Y.M. Dynamic Android Malware Analysis with De-Identification of Personal Identifiable Information. In Proceedings of the 2020 the 3rd International Conference on Computing and Big Data, Taichung, Taiwan, 5–7 August 2020; pp. 30–36. [Google Scholar]
  25. Chew, C.J.W.; Kumar, V.; Patros, P.; Malik, R. ESCAPADE: Encryption-Type-Ransomware: System Call Based Pattern Detection. In Proceedings of the International Conference on Network and System Security, Melbourne, Australia, 25–27 November 2020; Springer: New York, NY, USA, 2020; pp. 388–407. [Google Scholar]
  26. Mercaldo, F. A framework for supporting ransomware detection and prevention based on hybrid analysis. J. Comput. Virol. Hacking Tech. 2021, 17, 221–227. [Google Scholar] [CrossRef]
  27. El-Shafai, W.; Almomani, I.; AlKhayer, A. Visualized Malware Multi-Classification Framework Using Fine-Tuned CNN-Based Transfer Learning Models. Appl. Sci. 2021, 11, 6446. [Google Scholar] [CrossRef]
  28. Sharma, S.; Singh, S. Texture-Based Automated Classification of Ransomware. J. Inst. Eng. (India) Ser. B 2021, 102, 131–142. [Google Scholar] [CrossRef]
  29. Vasan, D.; Alazab, M.; Wassan, S.; Safaei, B.; Zheng, Q. Image-Based malware classification using ensemble of CNN architectures (IMCEC). Comput. Secur. 2020, 92, 101748. [Google Scholar] [CrossRef]
  30. Pinhero, A.; Anupama, M.; Vinod, P.; Visaggio, C.A.; Aneesh, N.; Abhijith, S.; AnanthaKrishnan, S. Malware Detection employed by Visualization and Deep Neural Network. Comput. Secur. 2021, 105, 102247. [Google Scholar] [CrossRef]
  31. Yuan, B.; Wang, J.; Liu, D.; Guo, W.; Wu, P.; Bao, X. Byte-level malware classification based on markov images and deep learning. Comput. Secur. 2020, 92, 101740. [Google Scholar] [CrossRef]
  32. Singh, J.; Sharma, K.; Wazid, M.; Das, A.K. SINN-RD: Spline interpolation-envisioned neural network-based ransomware detection scheme. Comput. Electr. Eng. 2023, 106, 108601. [Google Scholar] [CrossRef]
  33. Fernando, D.W.; Komninos, N. FeSA: Feature selection architecture for ransomware detection under concept drift. Comput. Secur. 2022, 116, 102659. [Google Scholar] [CrossRef]
  34. Alissa, K.A.; Elkamchouchi, D.H.; Tarmissi, K.; Yafoz, A.; Alsini, R.; Alghushairy, O.; Mohamed, A.; Al Duhayyim, M. Dwarf mongoose optimization with machine-learning-driven ransomware detection in internet of things environment. Appl. Sci. 2022, 12, 9513. [Google Scholar] [CrossRef]
  35. De Gaspari, F.; Hitaj, D.; Pagnotta, G.; De Carli, L.; Mancini, L.V. Evading behavioral classifiers: A comprehensive analysis on evading ransomware detection techniques. Neural Comput. Appl. 2022, 34, 12077–12096. [Google Scholar] [CrossRef]
  36. Herrera-Silva, J.A.; Hernández-Álvarez, M. Dynamic Feature Dataset for Ransomware Detection Using Machine Learning Algorithms. Sensors 2023, 23, 1053. [Google Scholar] [CrossRef]
  37. Kim, H.; Park, J.; Kwon, H.; Jang, K.; Seo, H. Convolutional Neural Network-Based Cryptography Ransomware Detection for Low-End Embedded Processors. Mathematics 2021, 9, 705. [Google Scholar] [CrossRef]
  38. Bello-Salau, H.; Onumanyi, A.; Salawudeen, A.; Mu’azu, M.; Oyinbo, A. An examination of different vision based approaches for road anomaly detection. In Proceedings of the 2019 2nd International Conference of the IEEE Nigeria Computer Chapter (NigeriaComputConf), Zaria, Nigeria, 14–17 October 2019; pp. 1–6. [Google Scholar]
  39. Almomani, I.; Alkhayer, A.; El-Shafai, W. An Automated Vision-Based Deep Learning Model for Efficient Detection of Android Malware Attacks. IEEE Access 2022, 10, 2700–2720. [Google Scholar] [CrossRef]
  40. Lee, J.; Lee, K. A method for neutralizing entropy measurement-based ransomware detection technologies using encoding algorithms. Entropy 2022, 24, 239. [Google Scholar] [CrossRef]
  41. Manavi, F.; Hamzeh, A. A novel approach for ransomware detection based on PE header using graph embedding. J. Comput. Virol. Hacking Tech. 2022, 18, 285–296. [Google Scholar] [CrossRef]
  42. Mahfouz, A.M.; Venugopal, D.; Shiva, S.G. Comparative analysis of ML classifiers for network intrusion detection. In Proceedings of the Fourth International Congress on Information and Communication Technology, London, UK, 25–26 February 2019; Springer: Singapore, 2020; pp. 193–207. [Google Scholar]
  43. Thakkar, A.; Lohiya, R. Attack classification using feature selection techniques: A comparative study. J. Ambient Intell. Humaniz. Comput. 2021, 12, 1249–1266. [Google Scholar] [CrossRef]
  44. Alshammri, G.H.; Samha, A.K.; Hemdan, E.E.D.; Amoon, M.; El-Shafai, W. An Efficient Intrusion Detection Framework in Software-Defined Networking for Cybersecurity Applications. CMC-Comput. Mater. Contin. 2022, 72, 3529–3548. [Google Scholar] [CrossRef]
  45. Ni, S.; Qian, Q.; Zhang, R. Malware identification using visualization images and deep learning. Comput. Secur. 2018, 77, 871–885. [Google Scholar] [CrossRef]
  46. Namanya, A.P.; Awan, I.U.; Disso, J.P.; Younas, M. Similarity hash based scoring of portable executable files for efficient malware detection in IoT. Future Gener. Comput. Syst. 2020, 110, 824–832. [Google Scholar] [CrossRef]
  47. Brownlee, J. Deep Learning with Python: Develop Deep Learning Models on Theano and TensorFlow Using Keras; 2016. Available online: https://www.udemy.com/course/deep-learning-with-python-and-keras/ (accessed on 1 January 2020).
  48. Hodnett, M.; Wiley, J.F. R Deep Learning Essentials: A Step-by-Step Guide to Building Deep Learning Models Using TensorFlow, Keras, and MXNet; 2018. Available online: https://www.amazon.com/Deep-Learning-Essentials-step-step/dp/178899289X (accessed on 1 January 2020).
  49. Vasilev, I.; Slater, D.; Spacagna, G.; Roelants, P.; Zocca, V. Python Deep Learning: Exploring Deep Learning Techniques and Neural Network Architectures with Pytorch, Keras, and TensorFlow; 2019. Available online: https://searchworks.stanford.edu/view/13246756 (accessed on 4 August 2020).
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  51. Hong, F.; Lu, C.; Jiang, W.; Ju, W.; Wang, T. RDNet: Regression Dense and Attention for Object Detection in Traffic Symbols. IEEE Sens. J. 2021, 21, 25372–25378. [Google Scholar] [CrossRef]
  52. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  53. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  54. Zhou, B.; Lapedriza, A.; Khosla, A.; Oliva, A.; Torralba, A. Places: A 10 million image database for scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 1452–1464. [Google Scholar] [CrossRef]
  55. Rezende, E.; Ruppert, G.; Carvalho, T.; Ramos, F.; De Geus, P. Malicious software classification using transfer learning of resnet-50 deep neural network. In Proceedings of the 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA), Cancun, Mexico, 18–21 December 2017; pp. 1011–1014. [Google Scholar]
  56. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  57. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  58. Almohaini, R.; Almomani, I.; AlKhayer, A. Hybrid-based analysis impact on ransomware detection for Android systems. Appl. Sci. 2021, 11, 10976. [Google Scholar] [CrossRef]
  59. Alsoghyer, S.; Almomani, I. Ransomware detection system for Android applications. Electronics 2019, 8, 868. [Google Scholar] [CrossRef]
Figure 1. Malware classification approaches.
Figure 1. Malware classification approaches.
Sensors 23 04467 g001
Figure 2. Structure of the proposed end-to-end ransomware detection system.
Figure 2. Structure of the proposed end-to-end ransomware detection system.
Sensors 23 04467 g002
Figure 3. Samples of visual color and gray images of ransomware and benign apps. (a) Visual images of ransomware apps. (b) Visual images of benign apps.
Figure 3. Samples of visual color and gray images of ransomware and benign apps. (a) Visual images of ransomware apps. (b) Visual images of benign apps.
Sensors 23 04467 g003
Figure 4. Structure of the fine-tuned ResNet50 model.
Figure 4. Structure of the fine-tuned ResNet50 model.
Sensors 23 04467 g004
Figure 5. Binary confusion matrix.
Figure 5. Binary confusion matrix.
Sensors 23 04467 g005
Figure 6. The obtained confusion matrix and ROC curve for the best accomplished ML-based AB detection model. (a) Confusion matrix. (b) ROC curve.
Figure 6. The obtained confusion matrix and ROC curve for the best accomplished ML-based AB detection model. (a) Confusion matrix. (b) ROC curve.
Sensors 23 04467 g006
Figure 7. The obtained accuracy and loss curves for the training and validation processes of the best accomplished ResNet50 classification model using the color image dataset.
Figure 7. The obtained accuracy and loss curves for the training and validation processes of the best accomplished ResNet50 classification model using the color image dataset.
Sensors 23 04467 g007
Figure 8. The obtained accuracy and loss curves for the training and validation processes of the best accomplished ResNet50 classification model using the gray image dataset.
Figure 8. The obtained accuracy and loss curves for the training and validation processes of the best accomplished ResNet50 classification model using the gray image dataset.
Sensors 23 04467 g008
Figure 9. The obtained confusion matrix for the best accomplished ResNet50 classification model using (a) color image dataset and (b) gray image dataset.
Figure 9. The obtained confusion matrix for the best accomplished ResNet50 classification model using (a) color image dataset and (b) gray image dataset.
Sensors 23 04467 g009
Figure 10. The obtained AROC curves for the best accomplished ResNet50 classification model using (a) color image dataset and (b) gray image dataset.
Figure 10. The obtained AROC curves for the best accomplished ResNet50 classification model using (a) color image dataset and (b) gray image dataset.
Sensors 23 04467 g010
Table 2. Typical image widths for different sizes of benign and ransomware APKs.
Table 2. Typical image widths for different sizes of benign and ransomware APKs.
APK SizeImage Width
<10 KB32
10∼30 KB64
30∼60 KB128
60∼100 KB256
100∼200 KB384
200∼500 KB512
500∼1000 KB768
1000∼2000 KB1024
2000∼4000 KB1280
4000∼8000 KB1536
8000 KB∼10 MB1792
10∼15 MB2048
15 ∼20 MB2560
20∼25 MB3072
25∼30 MB4096
>30 MB5120
Table 3. The examined CNN models and their input image resolutions.
Table 3. The examined CNN models and their input image resolutions.
No.Model NameImage Size
1ResNet50224 × 224
2AlexNet227 × 227
3InceptionV3299 × 299
4ResNet101224 × 224
5GoogleNet224 × 224
6VGG16224 × 224
7DarkNet53256 × 256
8Xception299 × 299
9InceptionResNetV2299 ×299
10MobileNetV2224 × 224
11NasNetMobile224 × 224
12DarkNet19256 × 256
13ResNet18224 × 224
14DenseNet201224 × 224
15NasNetLarge331 × 331
16Places365-GoogleNet224 × 224
17ShuffleNet224 × 224
18SqueezeNet227 × 227
19VGG19224 × 224
Table 4. Fine-tuning and training optimization parameters utilized in the proposed vision-based RD approach.
Table 4. Fine-tuning and training optimization parameters utilized in the proposed vision-based RD approach.
ParameterValue
Learning ratio (LR)0.00001
Optimization approachADAM
Regularization approachL2-regularizer
Regularization decay rate0.001
Number of epochs20
Minimum batch size16
Validation frequency16
Dropout rate0.5
LR schedule parameterPiecewise
LR drop period parameter3
LR drop factor parameter0.9
Loss functionCategorical cross-entropy
Shuffling scenarioPerformed every epoch
Table 5. The obtained evaluation metrics for the examined ML detection models.
Table 5. The obtained evaluation metrics for the examined ML detection models.
ML ModelAccuracyRecallPrecisionF1-Score
SVM0.9250.9250.9250.925
RF0.9450.9450.9450.94
AB0.970.970.970.97
CART0.940.940.940.94
KNN0.930.9290.930.93
NB0.880.880.8890.88
LDA0.900.900.900.90
LR0.840.840.840.84
Table 6. Performance analysis results of the examined CNN models on color image dataset.
Table 6. Performance analysis results of the examined CNN models on color image dataset.
ModelVal. Acc.Rec. (TPR)Prec. (PPV)NPVSpec. (TNR)F1-ScoreAROCFNRFPRFORFDRMis Class. Rate
ResNet500.99510.9910.990.9950.999800.0100.00990.005
AlexNet0.9950.9910.990110.99510.0100.009900.005
InceptionV30.9950.9910.990110.9950.9990.0100.009900.005
ResNet1010.9950.9910.990110.9950.99980.0100.009900.005
GoogleNet0.990.990.990.990.990.990.99770.010.010.010.010.01
VGG160.990.9810.980410.989910.0200.019600.01
DarkNet530.990.9810.980410.98990.990.0200.019600.01
Xception0.990.990.990.990.990.990.99810.010.010.010.010.01
InceptionResNetV20.9910.980410.980.990.998100.0200.01960.01
MobileNetV20.98510.970910.970.98520.999500.0300.02910.015
NasNetMobile0.9850.990.98020.98990.980.9850.99840.010.020.01010.01980.015
DarkNet190.9850.990.98020.98990.980.9850.99260.010.020.01010.01980.015
ResNet180.98510.970810.970.98520.996200.0300.02910.015
DenseNet2010.98510.970810.970.98520.998800.0300.02910.015
NasNetLarge0.9850.990.98020.98990.980.9850.99840.010.020.01010.01980.015
Places365-GoogleNet0.9810.961510.960.98040.999100.0400.03840.02
ShuffleNet0.980.970.98980.97060.990.97980.99040.030.010.02940.01020.02
SqueezeNet0.980.990.97060.98970.970.98020.99730.010.030.01020.02940.02
VGG190.970.970.970.970.970.970.9950.030.030.030.030.03
Table 7. Performance analysis results of the examined CNN models on gray image dataset.
Table 7. Performance analysis results of the examined CNN models on gray image dataset.
ModelVal. Acc.Rec. (TPR)Prec. (PPV)NPVSpec. (TNR)F1-ScoreAROCFNRFPRFORFDRMis Class. Rate
ResNet500.990.990.9910.990.990.99970.010.010.010.010.01
NasNetMobile0.9850.990.98020.98990.980.98510.99830.010.020.01010.01980.015
MobileNetV20.9850.9710.970910.98470.99120.0300.029100.015
InceptionV30.9850.9710.970910.98470.99990.0300.029100.015
GoogleNet0.9850.980.98990.98020.990.98490.99970.020.010.01980.01010.015
VGG160.980.9610.961510.97960.99960.0400.038500.02
AlexNet0.980.980.980.980.980.980.99390.020.020.020.020.02
DarkNet190.980.980.980.980.980.980.99250.020.020.020.020.02
Places365-GoogleNet0.97510.952310.950.97560.994200.0500.04760.025
ResNet180.97510.952310.950.97560.999600.0500.04760.025
ResNet1010.9750.970.97980.97030.980.97490.99490.030.020.02970.02020.025
DarkNet530.9750.980.97030.97980.970.97510.9990.020.030.02020.02970.025
ShuffleNet0.9750.970.97980.97030.980.97490.99860.030.020.02970.02020.025
Xception0.9750.960.98970.96120.990.97460.99920.040.010.03880.01030.025
InceptionResNetV20.9750.990.96120.98970.960.97540.99850.010.040.01030.03880.025
SqueezeNet0.970.960.97960.96080.980.96960.98710.040.020.03920.02040.03
DenseNet2010.970.960.97960.96080.980.96960.9980.040.020.03920.02040.03
VGG190.960.950.96940.95090.970.95960.99620.050.030.0490.03060.04
NasNetLarge0.960.990.9340.98940.930.96120.99430.010.070.01060.0660.04
Table 8. The storage size (gigabytes) of the color and gray images of the ransomware and benign samples.
Table 8. The storage size (gigabytes) of the color and gray images of the ransomware and benign samples.
Android AppsColor SamplesGray Samples
Ransomware0.3980.353
Benign3.83.79
Total4.1984.143
Additional storage size/incremental percentage produced by color samples0.055 GB/1.31%
Table 9. Specifications of the utilized FT CNN models.
Table 9. Specifications of the utilized FT CNN models.
CNN ModelDisk Size (MB)LayersParameters (Total)Parameters (Trained)Parameters (Non-Trained)Reduced (%) in Training Parameters
ResNet50965025,600,000409625,595,90499.98
AlexNet227861,000,000819260,991,80899.99
InceptionV3894823,900,000409623,895,90499.98
ResNet10116710144,600,000409644,595,90499.99
GoogleNet27227,000,00020486,997,95299.97
VGG1651516138,000,0008192137,991,80899.99
DarkNet531555341,600,000204841,597,95299.99
Xception857122,900,000409622,895,90499.98
InceptionResNetV220916455,900.00307255,896,92899.99
MobileNetV213533,500,00025603,497,44099.93
NasNetMobile20*5,300,00021125,297,88899.96
DarkNet19781920,800,000204820,797,95299.99
ResNet18441811,700,000102411,698,97699.99
DenseNet2017720120,000,000384019,996,16099.98
NasNetLarge332*88,900,000806488,891,93699.99
Places365-GoogleNet272261,000,000204860,997,95299.99
ShuffleNet5.4507,000,00010246,998,97699.99
SqueezeNet5.2181,240,00010241,238,97699.92
VGG1953519144,000,0008192143,991,80899.99
* The NasNetLarge and NasNetMobile models do not comprise a linear structure of CNN modules.
Table 10. Computational analysis of the utilized FT CNN models.
Table 10. Computational analysis of the utilized FT CNN models.
CNN ModelColor SamplesGray Samples
Total Execution Time (s)Average Time per Sample (s)Total Execution Time (s)Average Time per Sample (s)
ResNet5016171.6214991.5
AlexNet18901.8917141.71
InceptionV321662.1720502.05
ResNet10121832.1819391.94
GoogleNet14781.4814341.43
VGG1619661.9718031.8
DarkNet5319751.9818771.88
Xception20212.0219341.93
InceptionResNetV237923.7937123.71
MobileNetV216361.6416041.6
NasNetMobile47934.7946094.61
DarkNet1914001.413771.38
ResNet1813771.3812791.28
DenseNet20146834.6846234.62
NasNetLarge50595.0649674.97
Places365-GoogleNet14961.514101.41
ShuffleNet15691.5715551.56
SqueezeNet13131.3112481.25
VGG1922202.2221472.15
Table 11. Comparative analysis outcomes.
Table 11. Comparative analysis outcomes.
Detection SystemDetection Accuracy
Proposed (FT ResNet5)99.5
Proposed (Ada Boost)97
[1]94.5
[2]96.4
[58]64.8
[59]96.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Almomani, I.; Alkhayer, A.; El-Shafai, W. E2E-RDS: Efficient End-to-End Ransomware Detection System Based on Static-Based ML and Vision-Based DL Approaches. Sensors 2023, 23, 4467. https://doi.org/10.3390/s23094467

AMA Style

Almomani I, Alkhayer A, El-Shafai W. E2E-RDS: Efficient End-to-End Ransomware Detection System Based on Static-Based ML and Vision-Based DL Approaches. Sensors. 2023; 23(9):4467. https://doi.org/10.3390/s23094467

Chicago/Turabian Style

Almomani, Iman, Aala Alkhayer, and Walid El-Shafai. 2023. "E2E-RDS: Efficient End-to-End Ransomware Detection System Based on Static-Based ML and Vision-Based DL Approaches" Sensors 23, no. 9: 4467. https://doi.org/10.3390/s23094467

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop