Next Article in Journal
Analysis of the Temperature Field Effect on the Thermal Stress of the Main Tower of Long-Span Suspension Bridges
Next Article in Special Issue
Special Issue on Advanced Technology of Intelligent Control and Simulation Evaluation
Previous Article in Journal
Extreme Wave Loading on a Vertical Circular Cylinder
Previous Article in Special Issue
PID Control Model Based on Back Propagation Neural Network Optimized by Adversarial Learning-Based Grey Wolf Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Systematic Parameter Analysis of Cloud Simulation Tools in Cloud Computing Environments

by
Muhammad Asim Shahid
1,
Muhammad Mansoor Alam
1,2,3,4 and
Mazliham Mohd Su’ud
3,*
1
Malaysian Institute of Information Technology, Universiti Kuala Lumpur, Kuala Lumpur 50250, Malaysia
2
Faculty of Computing, Riphah International University, Sector I-14, Hajj Complex, Islamabad 46000, Pakistan
3
Persiaran Multimedia, Multimedia University, Cyberjaya 63100, Malaysia
4
School of Computer Science, University of Technology Sydney, 15 Broadway, Ultimo, NSW 2007, Australia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(15), 8785; https://doi.org/10.3390/app13158785
Submission received: 27 April 2023 / Revised: 29 May 2023 / Accepted: 29 June 2023 / Published: 29 July 2023
(This article belongs to the Special Issue Advanced Technology of Intelligent Control and Simulation Evaluation)

Abstract

:
To provide various applications in various domains, a large-scale cloud data center is required. Cloud computing enables access to nearly infinite computing resources on demand. As cloud computing grows in popularity, researchers in this field must conduct real-world experiments. Configuring and running these tests in an actual cloud environment is costly. Modeling and simulation methods, on the other hand, are acceptable solutions for emulating environments in cloud computing. This research paper reviewed several simulation tools specifically for cloud computing in the literature and presented the most effective simulation methods in this research domain, as well as an analysis of a variety of cloud simulation tools. Cloud computing tools such as CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports were evaluated. Furthermore, a parametric evaluation of cloud simulation tools is presented based on the identified parameters. Several 5-parameter tests were performed to demonstrate the capabilities of the cloud simulator. These results show the value of our proposed simulation system. CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports are used to evaluate host processing elements, virtual machine processing elements, cloudlet processing elements, userbase average, minimum, and maximum, and cloudlet ID Start Time, Finish Time, Average Start, and Average Finish for each simulator. The outcomes compare these five simulator metrics. After reading this paper, the reader will be able to compare popular simulators in terms of supported models, architecture, and high-level features. We performed a comparative analysis of several cloud simulators based on various parameters. The goal is to provide insights for each analysis given their features, functionalities, and guidelines on the way to researchers’ preferred tools.

1. Introduction

These days, smart devices have become prominent in supporting users in various ways. Smart devices have proven useful for the user regardless of location by using improved features and offering resources available wherever and whenever to support their needs [1]. Using hardware and software to provide a service over a network (typically the internet) is called CC. The CC is an Internet-based computing model that shares on-demand resources (e.g., networks, servers, storage, apps, and services), software, and information with different user devices [2]. Users can access data and use the software with cloud storage from any computer that can connect to the internet. Google’s Gmail is an example of a cloud service provider [3]. In CC, users can access resources through the internet all the time. The users only have to pay for those resources based on their use. The cloud provider of cloud processing outsources each resource belonging to the client [4]. Researchers in CC need to perform actual experiments in their studies, setup, implementation, and experiments in the cloud; they often have a massive cost of creating a real CE. Using models and simulation software to replicate CEs and perform prerequisite testing is one viable alternative. CloudSim, CloudSim Plus [5], CloudAnalyst [6], iFogSim, and CloudReports are some of the existing cloud simulation tools discussed in this study. As a result, it is critical to select an efficient LB tool that meets QoS requirements [5]. An in-depth study on cloud simulators has been conducted in this article that identifies several parameters to evaluate them [7].
Current work is based on the analysis of HPEs, VM PEs, cloudlet PEs, UBs Avg, Min, Max, and cloudlet ID Start Time, Finish Time, Average Start, and Average Finish for each simulator, and cloud simulation tool reports are used, which circularly handle the requests of the customer based on power usage of hosts and VM consumptions, CDC, SB, and CPU usage with the host. The customer directly forwards the request through the internet, and the SB assigns the load to the DC. The SB keeps a record of all available DCs and the next DC to whom the next task shall be assigned when a customer’s request is submitted through the internet, then it forwards the request to the SB, and then the SB circularly selects the DC and assigns the job to the DC [8]. The most extensively used platform, OpenStack, is attracting more and more attention as it seeks to be competitive with other platforms such as AWS. Security issues are prevalent throughout the lifecycle of OpenStack, making security analysis an essential task [9]. The paper contribution is shown below:
  • In this article, the setting of the simulators is carried out to examine the many metrics that the cloud client can measure, specifically the inbound and outbound HPE.
  • This study has used a variety of simulators and experiments that may have an impact on future cloud simulation studies.
  • The prominent simulators for experimental campaigns and analysis for productive work are mentioned in this study report.
  • Based on this study and to the best of our knowledge, there is no succeeding work associated with other simulators.

1.1. Motivation

  • Cloud computing is growing in popularity, and researchers need to conduct real-world experiments. Configuring and running these tests in an actual cloud environment is costly, so we need to remove these issues with the help of cloud simulators.
  • This paper provides popular cloud simulators for modeling cloud computing environments, which enable scheduling VMs and evaluating different parameters, makespans, and execution times to reduce the complexity of testing cloud systems.
  • The goal of this analysis is to make it easier for developers and researchers to analyze clouds and test different hypotheses in a controlled environment.

1.2. OpenStack Cloud Management Platform

With the help of NASA and Rackspace Hosting, OpenStack was created in 2010 as a production IaaS platform. In a DC, it controls the computing, storage, and networking resources as a dashboard that allows administrators management abilities while enabling users to provision resources using a web interface. It is possible to construct both private and public clouds using OpenStack. There are 22 versions of OpenStack, ranging from A (Austin) to V (Victoria). In a DC or a network of DCs, several sets of hypervisors, storage, and networking hardware are combined into resource pools by OpenStack. Additionally, OpenStack offers a cloud architecture that has enormous advantages in terms of robustness, performance, and compatibility when compared to cloud management platforms (Eucalyptus, Apache Cloud-Stack 4.18.0.0) [9].

1.3. Differences between OpenStack and Current Cloud Simulators

  • OpenStack is no exception, and security issues are present in its lifecycle [9].
  • Along with the OpenStack cloud, Ubuntu is available in a more advanced form [9].
  • All cloud apps are compatible with OpenStack [9].
  • Only a small number of resources are needed to experiment with OpenStack [9].
  • Cloud simulators can be entirely changed and are completely open-source [9].
  • Cloud simulators have unlimited support for testing [9].
  • A simple desktop machine is sufficient to work on cloud simulators [9].
  • Cloud simulators provide a graphical representation of results [9].
This paper consists of eight sections: The first section contains the introduction, and Section 2 is based on popular categories of cloud simulators for this research study. Related work regarding the research’s benefits and drawbacks is in Section 3. The research methodology and proposed work are described in Section 4. Section 5 is based on experimental results and findings that included a cloud simulator analysis, and Section 6 contains a discussion of this research study. Section 7 is based on the recommendations of real-world scenarios for the reader, and Section 8 concludes this study and makes recommendations for future work.

2. Highly Prominent and Extensible Categories of Cloud Simulators

Cloud simulators are based on two popular categories: (A) common cloud simulators and (B) provisioning datacenters. Common cloud simulators include CloudSim, CloudSim Plus, CloudAnalyst, and iFogSim. Provisioning DCs have energy-aware provisioning with a CloudReport simulator. Figure 1 demonstrates the popular categories of cloud simulators [10].

2.1. Prominent Cloud Simulators

These tools are easy to expand and extremely versatile to use. Thereafter, only the characteristics that are of concern are expanded, while those contradictions are neglected. The resources that are described in this segment fall under this group [10].

2.1.1. CloudSim

According to Mansouri et al. [11], CloudSim is a toolkit for modeling CEs proposed by the CLOUD Laboratory, University of Melbourne, Australia. CloudSim models DCs, VMs, SBs, and resource methods to model provisioning. It provides a flexible transition between space-shared and time-shared resource allocations for processing components. Researchers can deploy cloud-based apps in a variety of environments with minimal effort and time spent evaluating the results. According to Sundas and Narayan Panda [12], CloudSim provides simulation, smooth modeling, and functional design within the context of the CC. It is a forum for the design of SBs, DCs, and large-scale allocation policies and scheduling. It has some extended features, such as a virtualization engine, to design the development and monitoring of the lifecycle of this system in the DC. The GridSim architecture developed in the lab for power systems serves as the foundation for CloudSim’s infrastructure. Puhan et al. [13] proposed IaaS (Service Infrastructure) layer modeling for cloud facilitation DCs that can be defined in CloudSim. With CloudSim, one can build various models to explain energy consumption in CEs. Users can simulate VM migration, host machine termination, and energy models. According to Hassaan [14], CloudSim’s simulation framework is a Java-based simulation toolkit that can design extremely large-scale clouds and support communication among cloud system components, as suggested. It is an event-focused simulator that has a unique feature known as a federated strategy. A federated cloud is a collection of internal and external cloud services that are distributed to meet business needs. Ismail [15] suggested that CloudSim is a virtualization and multi-tenant infrastructure simulation environment. CloudSim’s architecture was inspired by the limitations of true testbeds (such as Amazon EC2) in terms of scaling and replication of test results. As a result, CloudSim provides a generalized and extensible simulation framework that includes modeling, simulation, and testing aspects, which is especially useful for large-scale DCs in the cloud. It enables a cost-free, reproducible, and manageable environment and adjusts for success bottlenecks before deployment on real clouds.

2.1.2. CloudSim Plus

According to Filho et al. [16], CloudSim Plus is a Java 8 simulation system that allows various CC services to be modeled and simulated, ranging from IaaS to SaaS. It enables the introduction of simulation scenarios for the research, evaluation, and validation of algorithms for a variety of purposes. CloudSim Plus is an open-source simulation framework that aims to provide a tool that is extendable, modular, and accurate by adhering to software engineering principles and an object-oriented architecture. According to Naik and Mehta [7], CloudSim Plus, an object-oriented simulation application that is an extension of CloudSim, was proposed as free software. The CloudSim Plus API is the main component representing the simulation system API and the use of cloud simulation experiments must be permitted according to the API. According to Markus and Kertesz [17], CloudSim Plus is a redesigned and rebranded version of CloudSim that aims to make simulation easier to use and more accurate. Additionally, it is available on GitHub.

2.1.3. CloudAnalyst

Anusooya et al. [18] reported that the CloudAnalyst simulation tool compares the output of different algorithms to the total response time. The CloudAnalyst tool looks uncomplicated, as it has a graphical environment that seems easy to apply. The visual outcome allows users to interpret findings more efficiently and readily. CloudAnalyst features include: (1) Being easy to apply; (2) defining simulation features with wide configuration points and flexibility; (3) generating results based on graphics; (4) allowing for experiment repetitions; and (5) having the capacity to store the output. Nandhini and Gnanasekaran [19] listed CloudAnalyst features such as: (1) Integrating cloud application analytics units into CloudSim; (2) geographically dispersing computer servers and client workloads for assessment using simulation in a large-scale cloud framework; (3) SBs maximizing the efficiency of software and service suppliers; and (4) the main units being composed of the datacenter, server, server broker, host, VM, and cloud coordinator. Desyatirikova and Khodar [20] indicated that CloudAnalyst would include a collection of elements that will provide the basis for CC, such as VMs, Cloudlets (Jobs and User Requests), DCs, SBs, and Hosts.

2.1.4. iFogSim

Bala and Chishti [21] suggest that iFogSim is a popular fog computing simulator. iFogSim has been established, and all CloudSim elements and features have been incorporated into iFogSim. The research community has overwhelmingly supported iFogSim, which can identify various performance parameters such as energy consumption, latency, response time, cost, and so on. It also allows for the replication of research in a stable environment. It employs a stream processing model in which sensors continuously generate data that fog devices evaluate, allowing users to quickly compare the results of various placement algorithms. Seo [22] suggested that iFogSim is one of the most popular CloudSim simulators. iFogSim can model IoT and fog conditions to calculate latency, network congestion, and device energy consumption. However, a simplified approach to modeling latency workloads is used. iFogSim will simulate a three-layer IoT fog system in various configurations. Although the simulator is used to test a variety of workloads, it fails in three-layer IoT fog systems because it fails to accurately model the computation schedule when processing items. It is well known that multi-core CPU efficiency interactions between device components do not expand sequentially. Markus and Kertesz [17] indicated that there are also more advanced iFogSim extensions. MyiFogSim (available on GitHub) helps mobile users handle VM migration. A key feature is the ability to replicate user mobility and its relationship to the VM migration policy. The iFogSimWithDataPlacement tool (available on GitHub) is another extension that proposes a method for data management studies into how data is stored in a fog system. For data location, it also analyzes latency, network utilization, and energy expenditure. Unfortunately, many simulators (such as Fog-Torch, OPNET, or SpanEdge) are designed to represent a single aspect of computing, making them less productive for general simulations.

2.2. Energy-Aware Provisioning

For the supplier, the energy supply is the most important because it can produce substantial savings. Initially, energy efficiency can be achieved both at the level of the network devices and at the level of the host. The normal parameter for estimating energy usage is primarily the processor fee. As this aspect is a significant contributor to the energy bill, more advanced simulators provide savings because they concentrate on modeling energy usage [10].

CloudReports

Zaidi [8] reported that CloudReports are a powerful and flexible simulation tool for computing DCs and resource energy using a GUI in a CC environment; with high-level data, it produces reports and organizes simulation results. The central enterprises of CloudReports are composed of clients, DCs, physical computers, VMs, network and network area space, and so on. These organizations work to run a simulation environment with the simulation administrator. Fairouz Fakhfakh et al. [23] proposed several changes to the top of the architecture of CloudSim. The proposal introduces a GUI that offers several features. First, it enables the simultaneous running of several simulations. With detailed information and the ability to export simulation data, it can produce reports. This data is linked to the usage costs of services, energy usage, ET, and so on.
Table 1 lists all of the key questions associated with the review’s subtopics, while Table 2 [24] shows the clear benefits and drawbacks of the reviewed simulators. Table 3 summarizes the key models of cloud simulation presented in this section I-A (CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports) for various parameters such as simulator type, simulation time, availability, hardware/software associated, programming language, basic policy, repetition of analysis, GUI, user expectations, and dynamic run. The energy-consuming parameters in the selected cloud simulators are listed in Table 4 [15].

3. Related Work

There has been a lot of research conducted on cloud simulation tools for CC. This review focuses on CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports. Table 5 summarizes the literature review, and Table 6 shows a comparative analysis of the literature review based on various factors.
According to Bahwaireth et al. [1], CloudSim, CloudAnalyst, CloudReports, CloudExp, GreenCloud, and iCanCloud are among the most effective simulation tools in the field of research. Some of these instruments are used to conduct experiments that demonstrate the cloud simulator’s capabilities. According to Mishra et al. [2], the simulation is carried out in the CloudSim simulator to analyze the performance of heuristic-based algorithms, and the results are presented. Rashid and Chaturvedi [3] tried to explore different CC services, apps, and features in this paper and provide unique examples of cloud services provided by the most popular cloud services, such as Google, Microsoft, and Amazon, with models of CC services and advantages. According to Prajapati and Sariya [4], load changes are spread across multiple hubs to ensure that no single hub is overburdened, which assists in the legitimate use of properties and also enhances the framework’s execution. Many current calculations offer stack changes and better use of properties. Bambrik [10] proposed that a detailed survey of the current CC simulators be undertaken to explore the characteristics, software design, and ingenuity behind these frameworks.
According to Motlhabane et al. [24], several cloud simulators have been developed over the years as cost-effective ways of performing cloud research tasks. Based on different measures, the authors conducted a comparative analysis of thirteen cloud simulators. Given the cloud simulator’s design, strengths, shortcomings, and recommendations, the aim was to provide insights into each of them to direct researchers on the choice of suitable instruments. T. Lucia Agnes Beena and Lawanya [34] introduced numerous cloud simulators that have been developed over the years to perform cloud testing tasks in an energy-efficient way. Almost ten cloud simulators based on energy-efficient parameters were evaluated in their paper, and a report was presented. Suryateja [35] suggested that different cloud simulators, which have been developed over the years, are a cost-effective way of performing tasks in cloud research. The paper compares cloud simulators based on different parameters and presents the findings and explanations that enable new researchers to select an acceptable cloud simulator. Khurana and Bawa [36] proposed that the quality metrics of cloud simulators be discussed and that the quality indicators addressed by them be integrated into each simulator. In the end, the authors conclude that there is a need to build a simulator that addresses acceptable quality metrics.
Sajjad et al. [37] suggested a comparative assessment of the most popular state-of-the-art energy-conscious cloud simulators, such as GreenCloud, CloudSim, iCanCloud, CloudAnalyst, NetworkCloudSim, and CloudReports. Khalil et al. [26] suggested that the common architecture of cloud simulators should be addressed and that cloud simulators should be evaluated based on various parameters. The outcomes were discussed and illustrated the various capabilities, problem suitability, and extensibility of the cloud simulators studied. Sumitha J. and Priya [38] proposed that CC has the principles of DCs, which are distributed or dispersed worldwide with single or multiple DCs, and virtualization over the internet with on-demand services. Shakir and Razzaque [39] suggested that a simulation be carried out using the CloudAnalyst framework to see the contrast of results between them. Among the frameworks being compared, the round-robin is the best.
Vashistha and Sholliya [40] suggested the availability of many open-source simulators for CC, such as CloudSim, CloudAnalyst, GreenCloud, iCanCloud, EMUSIM, GroudSim, and DCSim environments, and noted that a few principles of CC cannot be satisfactorily simulated by any of these simulators. Kumar et al. [41] conducted a basic and similar investigation of the current simulation tools in distributed computing. Based on the results of this study, further inquiries began to emerge. Malhotra [42] conducted a comparative analysis of current cloud simulators with the aim of helping end-users select the appropriate simulator. Maarouf et al. [43] investigated the different CC simulation platforms and tools and identified their strengths and weaknesses. Additionally, the paper notes the challenges and future directions for research. Pagare and Koli [44] propose an expandable simulation system called CloudSim that supports simulation, CC modeling, the development of one or more VMs on a simulated DC node, jobs, and the mapping to appropriate VMs, service management capacity, and cloud infrastructure modeling. CloudSim can also manage cloud storage mechanisms and enable the simulation of several DCs.
Gupta and Beri [45] suggested that the conceptual cost of purchasing the services of different providers could lead to an increase in money and time budgets or wastage. The solution to this problem is to test out the numerous simulation tools that are available on the market. The article lists some of the simulation methods used for simulation and modeling purposes. Asir Antony Singh et al. [46] proposed using simulation methods to test CC performance before building the cloud. They used CloudSim as a platform for modeling and simulating the CC world. Andrade and Nogueira [47] proposed the Stochastic Petri Net (SPN) approach to evaluate cloud-based DR solutions for IT environments. This approach enables different performance metrics to be evaluated (e.g., response time, throughput, availability, and others) and can therefore assist DR coordinators in choosing the most suitable DR solution and also validate the analytical approach’s accuracy by comparing analytical results with those obtained from the CloudSim Plus cloud simulator. Dogra and Singh [27] demonstrate the capabilities of the three well-known cloud simulators: (a) CloudSim, (b) CloudAnalyst, and (c) CloudReport. As with other cloud simulations, the difference is predicated on capacity and flexibility, as well as nearly identical systems and functionalities. These cloud simulators are software-based simulation tools that are assessed to determine their capabilities depending on a variety of parameters. An experimental evaluation of this equipment has been performed to determine the optimal method for cloud simulation. Ahmed and Zaman [48] reported that CloudSim Plus has a new, up-to-date, fully-featured, and fully-recorded simulation structure. CC infrastructure and application services are easy to use and develop, allowing for modeling, simulation, and experimentation. This enables developers to concentrate on particular system design problems to be explored without taking into consideration the low-level knowledge relevant to cloud-based infrastructures and services.
Gupta et al. [29] proposed iFogSim as an API based on the jJava programming language that inherits CloudSim’s existing API to handle its underlying discrete event-based simulation and for effective network-related workload handling. Researchers can first understand the project structure using CloudSim’s API before implementing it in iFogSim. Jena et al. [31] suggested the basic differences between cloud-based simulators in this research paper: CloudSim, CloudAnalyst, CloudReports, CloudSched, and GreenCloud. Based on the following parameters, such as extension, correlation, and classification, researchers carried out an extensive investigation with different groups and also took some of the essential open-source cloud resources for comparative analysis for better utility. Esparcia and Singh [32] suggested comprehensively exploring cloud simulation tools. Specifically, because each simulator has its own purpose, it is to suggest which simulator can fit one’s preferences. The four traditional and modern tools for modeling and studying the true cloud were discussed, and the simulation models, features, and simulation results of each of the simulation tools were also identified. Ashalatha R. et al. [33] described the popular simulation tools for large CEs that take care of computing resources and user workloads. Depending on the specifications, a variety of simulators can perform different simulations in the CE. Bhatia and Sharma [49] conducted a critical study and analysis of the different simulators available in the computer world for CC. The analysis will serve as an excellent source of reference for cloud simulators that might be required by researchers and academics. Results show that CloudSim is the core of many simulation algorithms for CC and is commonly used by most researchers. Simulators can perform different simulations in the CE. Shahid et al. [50] proposed that existing load-balancing techniques also explore the challenges of LB in the CC environment and identify the necessity for a novel load-balancing algorithm that uses FT metrics. Shahid et al. [51] proposed that the major goal of LB is to efficiently control load across numerous cloud nodes such that a node is under or overloaded. LB can help save time and money and improve the performance of a system by optimizing the entire system’s performance. Najat Tissir et al. [9] investigated the current state of the components, subcomponents, and interactions with OpenStack. The examination of the most prevalent OpenStack vulnerabilities is therefore the focus of their study based on ten years’ worth of security reports analysis. Alberto Nunez et al. [52] suggested using Cloud Expert, an intelligent system built on metamorphic testing that chooses the best simulator to cover the user’s desired characteristics and also analyzes the fundamental characteristics of some well-known cloud simulators to produce metamorphic rules that are then used to represent the simulator’s qualities. Harvinder Singh et al. [53] suggested that six cloud simulators be examined to develop and test various cloud apps. The study discusses various cloud simulators and contrasts them based on many factors, such as programming languages, support for availability and SLAs, and so forth. Youssef Saadi and Said El Kafhali [54] proposed an energy-efficient strategy (EES) for VM consolidation in a CE to lower energy consumption and increase job throughput. The performance-to-power ratio is used in the proposal to establish upper criteria for overload detection. Additionally, EES takes into account the total workload usage in the DC when establishing lower criteria, which can reduce the number of VM migrations. Daniel G. Lago et al. [55] suggested the use of SinergyCloud to assess DCs in hybrid clouds with a precise level of abstraction that enables the evaluation of various cloud situations, such as energy consumption, workflow makespan, the complete duration of tasks, and VM migrations. The simulation of hybrid clouds with numerous DCs and millions of devices is possible with SinergyCloud. It also has a less challenging learning curve than other simulators because it is a Java-based, event-driven, and packet-level simulator.
Table 5. The summary of the related work.
Table 5. The summary of the related work.
RefNameYearSimulatorAdvantagesDisadvantages
[1]Khadijah Bahwaireth et al.2016
  • CloudSim
  • CloudAnalyst
  • CloudReports
  • CloudExp
  • GreenCloud
  • iCanCloud
  • Improving capacity for data storage [7].
  • Improving reliability and protection through the collection of data and software on several servers.
  • When the main goal is to replicate distributed apps over several DCs and user groups, CloudAnalyst comes in handy.
  • A simple network model with only a minimal workload traffic generator.
  • Lack of modeling support for parallel experiments.
  • Lack of certain important components of CC, such as BPM and SLA.
[2]Sambit Kumar Mishra et al.2018CloudSim
  • Explanation of multiple performance parameters.
  • The calculation of the system’s makespan and the energy consumption is explained.
Not assessing the suggested algorithms for cloud deployment in the real world.
[3]Aaqib Rashid and Amit Chaturvedi2019CloudSimWith a broad range of apps, it offers cost-effective services.
  • Not addressing the problem of security in CC.
  • Lack of CC-related risk mitigation.
[4]Priyanka Prajapati and Amit Kumar Sariya2019CloudAnalystProvides flexibility, better use of assets, and better response times.
  • Lack of distributed storage.
  • Lack of stack adjustment.
  • In an IaaS CE, there is no QoS-based burden change system for various job activities.
[10]Ilyas Bambrik2020
  • CloudSim
  • CloudAnalyst
  • CloudReports
  • CloudExp
  • GreenCloud
  • Reproducibility of the outcome.
  • Cost-effectiveness.
  • Flexibility.
  • Unable to start running on a single host.
  • CloudSim does not provide the VM image transmission time and assumes that the VM will start immediately after the request is received.
[11]N. Mansouri et al.2020
  • CDOSim
  • TeachCloud
  • DartCSIM
  • DartCSIM+
  • Provides a comparative study of CE tools.
  • Provides recommendations on the selection of the most effective method for researchers.
  • Lack of security aspects in the cloud.
  • Not simple and customizable to use.
  • Inadequate scalability of the latest simulators.
  • Lack of distributed execution tool support.
  • The communication model is limited.
  • Lack of a network topology model, congestion control, and traffic patterns.
[12]Amit Sundas and Surya Narayan Panda2020CloudSimProvides simulation capability for performing several duplications at the same interval.
  • Lack of management for SLA.
  • Lack of automatic scaling of vertical VMs.
[13]Sharmistha Puhan et al.2020
  • SimGrid
  • GroudSim
  • GSSIM
  • GreenCloud
  • CloudSim
  • iCanCloud
  • Provides powerful schemes for energy.
  • Reduces the use of resources and improves QoS.
With fewer machines, the DC does not have an improved response time, and higher energy is consumed for higher-frequency memory configurations.
[14]Muhammad Hassaan2020
  • CloudSim
  • CloudAnalyst
  • GreenCloud
  • CloudSim and CloudAnalyst are Java-based simulation programs.
  • Consumption of resources according to abilities and capacity.
  • Provides various power-saving algorithms and descriptions of packet-level capture.
  • Lack of small DC activities.
  • CloudSim and CloudAnalyst lack power-saving modes.
  • CloudSim lacks a graphical output.
  • CloudSim and CloudAnalyst lack CPU simulation, memory, storage, and networking resources.
[15]Azlan Ismail2020
  • CloudSim
  • MDCSim
  • GDCSim
  • GreenCloud
  • DCSim
  • BigHouse
  • E-mc2
  • DCworms
  • CloudNetSim++
  • Provides support for energy-driven simulation.
  • CloudSim is a generalized and adaptable simulation platform that addresses modeling, simulation, and testing for large-scale cloud DCs.
  • Lack of holistic perspective in cloud simulation.
  • The number of real testbeds available in CloudSim is limited (e.g., Amazon EC2).
  • Providing comprehensive energy consumption equipment is still insufficient.
  • There are no memory management properties [15].
[16]Manoel C. Silva Filho et al.2017CloudSim Plus
  • CloudSim provides maintainability, reusability, extensibility, and accuracy.
  • CloudSim Plus contains unique characteristics, such as increased accuracy, expansion capability, and ease of use.
  • Lack of management for SLA.
  • Lack of automatic scaling of vertical VMs.
[7]Niti Naik and Mayuri A. Mehta2018
  • CloudSim
  • CloudAnalyst
  • GreenCloud
  • MDCSim
  • iCanCloud
  • ATAC4Cloud
  • CEPSim
  • CloudNetSim++
  • CloudSim Plus
Provides a catalog of cloud simulators available to assist researchers and developers in choosing the most suitable method for cloud simulation.
  • To use the simulator, the user must be familiar with the C++/Java language.
[17]Andras Markus and Attila Kertesz2020iFogSim
  • Easy network architecture.
  • Easier level of abstraction.
  • Lack of cost modeling.
  • Lack of control over energy-aware capabilities.
  • There are no VM management properties in this simulator.
  • Some critical cloud/fog simulation capabilities are missing.
[18]G. Anusooya et al.2020CloudSimIt offers cloud energy, power savings, and DC workload management.Lack of consistency and performance in simulation.
[19]J.M.Nandhini and T.Gnanasekaran2019CloudSimHelp in measuring costs and the use of assets.Insufficient storage problems and VM failures with variable ability.
[20]Elena N. Desyatirikova and Almothana Khodar2019CloudAnalyst
  • Providing optimal time for response.
  • Providing better distribution center efficiency.
  • Lack of current load from the DC.
  • Lack of availability of bandwidth.
[21]Mohammad Irfan Bala and Mohammad Ahsan Chishti2020iFogSimBy distributing the application modules between fog devices and cloud DCs, efficient utilization of the Cloud-Fog resources is achieved.There is a need for further improvements in algorithms for LB.
[22]Dongjoo Seo et al.2020iFogSimIn this framework, the controller module is integrated.
  • Lack of modeling for system reliability in iFogSim.
[8]Taskeen Zaidi2020CloudReports
  • Effective use of power within DCs.
  • CloudReports is a simulation tool with a GUI that creates reports and has an API with extensions and user-defined plugins.
  • Lack of Cloud Infrastructure Validation.
  • The configuration of large-scale DCs is extremely complicated because of the need to optimize resource utilization and power consumption.
[23]Fairouz Fakhfakh et al.2017
  • NetworkCloudSim
  • FederatedCloudSim
  • DynamicCloudSim
  • TeachCloudSim
  • CloudAnalyst
  • FTCloudSim
  • WorkFlowSim
  • ElasticSim
  • CloudReports
  • Providing a dynamic model for apps.
  • Different clouds (public, private, hybrid, and multi-cloud settings) are represented by CloudSim.
  • The constraint of time.
  • The current simulators do not allow for dynamic changes in execution based on certain rules.
[24]Neo Motlhabane et al.2018
  • CloudSim
  • CloudAnalyst
  • GreenCloud
  • iCanCloud
  • NetworkCloudSim
  • EMUSim
  • GroudSim
  • MR-CloudSim
  • SmartSim
  • DcSim
  • SimIC
  • Helps for SLA.
  • Availability.
  • Graphical interface.
  • Lack of a communication model.
  • Lack of a cost model.
  • CloudSim has no GUI and a limited communication prototype that does not recognize TCP/IP.
[34]T. Lucia Agnes Beena and J. Jenifer Lawanya2018
  • CloudSim
  • CloudAnalyst
  • Energy-efficient.
  • Developers can simulate and evaluate heterogeneous CEs with CloudSim and CloudAnalyst (Amazon EC2, Microsoft Azure).
There is no GUI in CloudSim.
[35]Pericherla S Suryateja2016
  • CloudSim
  • CloudAnalyst
  • In CC systems, the methods for obtaining the energy consumption of each hardware component are provided.
  • CloudSim is built on Java, so it is simple to add new modules. CloudSim is free to modify and enhance because it is open-source.
  • Advanced security techniques for VMs, such as privacy, integrity, and encryption, are not supported.
  • One major flaw in CloudSim is the lack of a GUI.
[36]Ravi Khurana and Rajesh Kumar Bawa2016CloudSim
  • Reduce costs for deployment.
  • Fast check for results.
  • Provide flexibility and scalability.
  • Usage for small DCs only.
  • In CloudSim, the simulation results are presented in plain text format.
[25]James Byrne et al.2017
  • CACTOSim
  • CDOSim
  • CEPSim
  • Cloud2Sim
  • CloudAnalyst
  • CloudExp
  • CloudNetSim++
  • The de facto base platform.
  • CloudSim and CloudReports have platform portability.
  • Lack of execution is distributed.
  • CloudSim and CloudReports do not have a web API.
  • There is no such thing as distributed architecture.
  • There is no web API available.
  • Only text output is available.
[37]Ammara Sajjad et al.2018
  • GreenCloud
  • CloudSim
  • iCanCloud
  • CloudAnalyst
  • NetworkCloudSim
  • CloudReports
  • Evaluate the implementation costs.
  • Precise estimation of energy consumption.
  • CloudAnalyst is available for Linux, Windows, and Mac computers.
  • Time-consuming task.
  • For performing simulations, CloudSim provides a restricted energy-based profile.
  • TCP/IP implementation is not available in CloudAnalyst, and power consumption support is limited.
[26]Khaled M. Khalil et al.2017
  • CloudSim
  • CloudAnalyst
  • GreenCloud
  • iCanCloud
  • MDCSim
CloudSim and CloudAnalyst are cost-effective simulators.
  • Reconfiguring and rescaling apps and hardware.
  • CloudReports has no power-saving mode.
[56]Soumya Ranjan Jena et al.2020CloudAnalystIn comparison to the other two models, Space-Shared and Dynamic-Workload, it uses fewer resources (CPU, RAM, and bandwidth).
  • Lack of QoS.
  • In comparison to TimeShared, Space-Share, and Dynamic-Workload, it utilizes a lot of resources (CPU, RAM, and bandwidth).
[38]Sumitha. J, and S.Manju Priya2018
  • CloudSim
  • CloudAnalyst
  • Provides scalable environments.
  • Distributed approach.
  • CloudSim is cost-beneficial.
  • CloudAnalyst produces more accurate simulation results.
  • Violations of information.
  • User information exploitation.
[39]Muhammad Sohaib Shakir, and Abdul Razzaque2017CloudAnalyst
  • Effectively access the assets.
  • CloudAnalyst is a performance analysis tool with a user-friendly GUI.
  • Degraded optimization of performance.
  • Poor response time.
[40]Avneesh Vashistha and Shourabh Sholliya2017CloudAnalyst
  • Simulation time is within seconds.
  • CloudAnalyst provides a GUI interface.
  • Lack of a communication model.
  • Because CloudSim is not graphical, researchers can use CloudAnalyst to get better visualization results.
[41]K. Arun Kumar et al.2017CloudReports
  • Controlled effectively.
  • CloudReports provides a GUI with a variety of features.
  • Lack of a suitable testing method for evaluating them.
  • There is no GUI in CloudSim.
[42]Manisha Malhotra2017
  • CloudSim
  • CloudAnalyst
  • CloudAnalyst provides the provisioning of services.
  • CloudAnalyst is user-friendly.
The resource provisioning mechanisms in CloudSim are limited, and it lacks a GUI.
[43]Adil Maarouf et al.2015CloudSimThe transition between space-shared and time-shared assignment of processing cores to virtualized services provides great versatility.
  • Provisioning of VM hosts.
  • CloudSim does not offer a ready-to-use environment for running a whole scenario with specific input.
[44]Jayshri Damodar Pagare and Nitin A Koli2015CloudSim
  • Energy consumption savings.
  • Flexible, scalable, reliable, and repeatable provisioning policy assessment for various apps.
  • The AVG usage of cloud DC services must be increased.
  • CloudSim does not provide energy efficiency services.
[45]Kiran Gupta and Rydhm Beri2016CloudSim
  • Simulation tools that have no investment in capital are involved.
  • Provides better results.
  • Server and DC hardware-related problems.
  • CloudSim is not providing full resources for networking.
[46]Asir Antony Gnana Singh et al.2018CloudSimGet the system’s ET under different CC organizations.Lack of performance.
[47]Ermeson Andrade and Bruno Nogueira2018CloudSim Plus
  • CloudSim Plus has a better response time and better accuracy.
  • No large-scale environment.
  • CloudSim Plus does not support hierarchical modeling, which would help minimize modeling complexity.
[27]Shobhna Dogra and A.J Singh2020
  • CloudSim
  • CloudAnalyst
  • CloudReports
  • Extensibility offers.
  • CloudAnalyst is a CloudSim-based application for modeling and analyzing large-scale CC environments.
  • On the subject of energy-aware CC infrastructures, CloudReports is a useful tool.
  • Problems with results and protection.
  • CloudSim is not providing PaaS and SaaS CEs.
  • CloudAnalyst has no migration policy.
[16]Manoelcampos2020CloudSim Plus
  • It is simpler to use.
  • Mechanism of Host Fault Injection.
Specific system design.
[29]Harshit Gupta2017
  • CloudSim
  • iFogSim
  • Minimal cost.
  • Huge potential for simulating the outcome.
Lack of inadequate infrastructure.
[28]David Abreu Perez et al.2019iFogSimBest ET.
  • Restricted to simulators in the cloud.
  • Only tree topologies are permitted.
  • Communication is only possible within the same branch of the tree.
[31]Soumya Ranjan Jena et al.2020
  • CloudSim
  • CloudAnalyst
  • CloudReports
  • CloudSched
  • GreenCloud
  • Open-source cloud resources.
  • Good design capabilities.
  • Components view.
  • Method of reproduction.
None of them is suitable for all experiences and agreements.
[32]Jay Ar P. Esparcia and Monisha Singh2017CloudAnalystExtensible, scalable, versatile, fast, open-source, user-friendly, and result-oriented resources are available.
  • Large procurement-related infrastructure costs.
  • High energy costs during implementation.
  • Limited usability of geographically located cloud services.
[33]Ashalatha R. et al.2016
  • SPECI
  • GreenCloud
  • iCanCloud
  • NetworkCloudSim
  • CDOSim
  • TeachCloud
  • Large environments in clouds.
  • Takes control of resources for storage and user workloads.
The simulation framework does not provide a federated CE.
[49]Mandeep S. Bhatia and Manmohan Sharma2016
  • GreenCloud
  • iCanCloud
  • CloudSim
  • DCSim
  • MDCSim
  • Freely accessible.
  • Extremely useful to researchers.
Selecting one over many others is a tough job.
[50]Muhammad Asim Shahid et al.2020CloudAnalystMeet the deadline for the makespan.Unable to predict when the burst will occur.
[51]Muhammad Asim Shahid et al.2021CloudSimFT is a feature of load-balancing algorithms, which means the researchers can provide standardized LB despite arbitrary node or connection errors.Lack of availability of key resources, as well as the installation of apps, is a concern.
[9]Najat Tissir et al.2020OpenStackIdentify the tendencies associated with its vulnerability.It can fail due to security issues.
[52]Alberto Nunez et al.2022Cloud Expert
  • Provided metamorphic testing.
  • Working on different aspects (energy, storage, network, memory, CPU).
Lack of graphical representation of results.
[53]Harvinder Singh et al.2021
  • GreenCloud
  • CloudSim
  • iCanCloud
  • CloudAnalyst
  • NetworkCloudSim
  • CloudReports
They provided effective simulators.Lack of providing a view of all simulators.
[54]Youssef Saadi and Said El Kafhali2020Energy-Efficient Simulator
  • Reducing energy consumption.
  • Provided the highest throughput.
Lack of real-time CE for variable factors.
[55]Daniel G. Lago et al.2021SinergyCloud
  • Will evaluate DCs in hybrid clouds.
  • Timely task completion for VMs.
  • Lack of native support for real-world workload
  • Lack of SLA violations control.
Table 6. A comparative analysis of the literature review based on various factors.
Table 6. A comparative analysis of the literature review based on various factors.
RefData StorageQOSCost-
Effectiveness
Network ModelSecuritySLAGUIFlexibilityAvailabilityReliability
[1]
[2]
[3]
[4]
[10]
[11]
[12]
[13]
[14]
[15]
[16]
[7]
[17]
[18]
[19]
[20]
[21]
[22]
[8]
[23]
[24]
[34]
[35]
[36]
[25]
[37]
[26]
[56]
[38]
[39]
[40]
[41]
[42]
[43]
[44]
[45]
[46]
[47]
[27]
[48]
[29]
[28]
[31]
[32]
[33]
[49]
[50]
[51]

4. Problem Statement

Because it is difficult to test new mechanisms in a real cloud computing environment and because researchers frequently cannot access a real cloud computing environment, using a simulation to model the mechanism and evaluate the results is required. Simulating a data center saves time and effort spent configuring a real testing environment.

5. Research Methodology

This section focuses on the cloud simulator architectures and proposes a selection strategy for the simulator based on research gaps. The architecture of the cloud simulators is incorporated and explained.

5.1. CloudSim Architecture

CloudSim was proposed in 2009 [43]. Figure 2 depicts the multifaceted software system scheme and structural components of CloudSim. The CloudSim simulation layer allows for the creation and simulation of virtualized cloud-based [24] and DC [38] environments with clear organizational boundaries for VMs, memory, storage, and bandwidth. This layer’s critical issues include, for example, assigning hosts to VMs, controlling request execution, and observing robust method states. The layer cost for memory and space use is reasonable [24]. There is no GUI [40], and it has a small prototype declaration that does not fund TCP/IP [24]. The CloudSim layer handled the initialization and operation of main identities (VMs, hosts, DCs, and apps) during the simulation test [44]. Many library functions written in the Java programming language are included in CloudSim [46].

5.2. CloudSim Plus Architecture

CloudSim Plus is a mix of various components. The principal component representing the simulation system API is the CloudSim Plus API. It is the single module needed to allow cloud simulation experiments to be implemented [16]. For the modeling and simulation of CC infrastructures and facilities, CloudSim Plus is an extensible simulation platform [47]. A condensed overview of its package structure is shown in Figure 3. Packages with a darker color include CloudSim Plus’s unique functionality. The most important CloudSim Plus bundles and classes are listed below [16]:
  • Distributions: Groups that produce pseudo-random numbers after many statistical distributions that are also used by the simulation API and the developers creating the simulations [16].
  • Network: Classes for building internet infrastructure for data centers that enable network simulations [16].
  • Power: Groups that allow power-aware simulations, namely models of energy usage that developer can expand in the simulations [16].

5.3. CloudAnalyst Architecture

The CloudAnalyst architecture is shown in Figure 4. CloudAnalyst offers an [35] easy-to-use GUI [39] to manage any geographically dispersed device, such as application workload definition, user location information, creating traffic, DC location, number of customers per DC, and number of DC resources. In the form of a map or table, CloudAnalyst may produce output that summarizes the massive number of users and device states during the simulation period [35]. It distinguishes between the environment for programming and the environment for simulation [45].

5.4. iFogSim Architecture

The iFogSim framework contains numerous components: (1) the iFogSim simulator expanded; (2) a Sniper Multi-Core simulator; (3) a System Cross-layer Controller; and (4) an Environment co-simulation module (see Figure 5). The Simulation Module Co-simulation Environment starts the simulation by configuring the loading device and application codes. Machine settings include hardware requirements for the virtual computer, the network layer, and the simulation framework. Program codes are the target programs that should be run on VMs. To begin with, the Sniper simulator runs the loaded app code on the specified system hardware configuration to generate statistics for subsequent moves. iFogSim uses the generated statistics to simulate the device and network layers based on the device specifications. The framework logs the ET and energy consumption of the application for the specified device configuration [22].

5.5. CloudReports Architecture

The CloudReports architecture is shown in Figure 6. It is an expandable simulation software for energy-aware CC environments that enables researchers to design several complicated concepts through an interface. On top of the [37] CloudSim simulation engine [56], it offers four layers: (1) Report manager; (2) Simulation manager; (3) Extensions; and (4) Core entities. The key benefit of CloudReports is its modular architecture, which enables its API to be expanded to experiment with new algorithms for scheduling and provisioning [25]. CloudReports supports modeling for restricted [37] power consumption [57] and is based on the mechanism of event-oriented simulation. CloudReports’ important features involve high accuracy and adaptive simulations based on help for auto-VM migration. It provides the Java Reflection API to build plug-in extensions [37]. It provides a GUI that offers unique characteristics [41]. Table 7 summarizes the various cloud simulator architectures.

5.6. Proposed Work with Selection Strategy for the Simulator Based on Research Gaps

Based on an extensive literature review, it has been found that most researchers have not considered experiments in a real environment, hence the need for a systematic analysis of cloud simulation tools.
This section proposes to evaluate cloud simulators for public clouds in particular. Only parameters that can be measured by the cloud client are used in the simulator’s configuration: inbound and outbound HPEs, VM PEs, cloudlet PEs, UB, and cloudlet ET parameters.
Various simulators and experiments were carried out in this study, which may have an impact on future cloud simulation work. The selected cloud simulators highlight the related succeeding works in Section 6 below. The following works are primarily based on the following parameters: HPEs, VM PEs, Cloudlet PEs, Avg, Min, Max, and Cloudlet ET. This research paper discusses popular simulators for experimental campaigns and successful work analysis. Based on this research and to the best of our knowledge, no successful work has been associated with other simulators. Figure 7 shows the framework for analysis and comparison.

6. Experimental Results and Findings

In this section, the paper’s contribution will be compared with the state-of-the-art simulators currently used in CC. With a focus on the increasing popularity of CC, the experiments of cloud simulators are added, and researchers in this field need to perform real experiments in the cloud simulator studies. It is expensive to configure and run these tests in actual CEs. Modeling and simulation methods, however, are sufficient solutions for these and also provide excellent alternatives for emulating environments in CC.
Throughout this section, Section 6.1Section 6.5 are used for source code and to evaluate the performance of changed CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports exploiting the power usage of hosts and VMs, the usage of the round-robin VM allocation policy, the process flow of the DC using a closet DC SB technique, and the round-robin DC broker policy with host and VM power consumption to capture a more realistic scenario where performance may change due to the power usage of hosts and VMs.

6.1. CloudSim

The following scenario evaluates CloudSim’s modeling and simulation capabilities. The goal of this scenario is to figure out the power usage of hosts and VMs. The simulation environment is composed of Scheduling Interval 10, Hosts 2, Hosts PEs (PEs) 8, VMs 4, VMs PEs 4, Cloudlets 8, Cloudlets PEs 2, and cloudlet length 50,000 for implementation purposes.
As per FCFS’s default policy, each VM is assigned to the first available host with available PEs. Each host distributes/shares its core capacity dynamically among all VMs running on the same time-shared VM scheduler (Pes). CloudSim’s modeling and simulation capabilities are depicted in Figure 8. Figure 8 shows that VMs are executing successfully according to the parameters that were given at the time of creation. This means that CloudSim is the best simulator for the power usage of hosts and VMs.

6.2. CloudSim Plus

The following scenario was built to test the modeling and simulation capabilities of CloudSim Plus and determine the use of the round-robin VM allocation policy. The simulation environment is composed of Hosts 4, Hosts PEs 8, VMs 8, VMs PEs 2, Cloudlets 8, Cloudlets PEs 2, and cloudlet length 10,000.
In this manner, it loads a VM onto a host and proceeds to the next host. Not all hosts are turned on when they are created. When VMs are added, hosts are activated on demand, as evidenced by logs. Remember that such a policy is extremely naive and increases the number of active hosts, resulting in higher power consumption. Figure 9 depicts the CloudSim Plus simulation workflow. In this experiment, the round-robin algorithm was integrated into CloudSim Plus to manage the load on the VM host. Once the first VM host has been loaded, this round-robin algorithm will automatically divide the load between VM hosts, which will be beneficial for cloud users in terms of time savings and cost.

6.3. CloudAnalyst

Figure 10, Figure 11, Figure 12 and Figure 13 show the workflow of the CloudAnalyst simulation method. In this experiment, we have six UBs for three DCs, which means that there are two UBs on one DC to check the response time of DCs. According to the result, DC 1 and DC 2 have five VMs to execute the cloud user tasks, and DC 3 has 50 VMs.
We pretend that an internet application is running in three different DCs around the world in this experiment. The service proximity policy is used to route user traffic to the DC that has the shortest network latency. This simulation will take approximately one hour to complete.
CloudAnalyst displays the regions with a response to time results when the simulation is completed, as depicted in Figure 11.
Figure 13 depicts the cloud infrastructure, which includes six geographically dispersed UBs and three DCs with 5 to 50 VMs.

6.4. iFogSim

The scenario below was created to add iFogSim’s modeling and simulation capabilities to the test. The amount of CPU time consumed by the host is determined by this scenario. The simulation environment consists of four VMs, some hosts (PEs) equal to the number of VMs, and VMs (PEs) for implementation purposes.
The iFogSim analysis is based on a DC with one host that places two VMs to run one cloudlet each and receives notifications when a host is allocated or reallocated to each VM. The example uses the new VM listeners to get these notifications while the simulation is running. It also shows how to reuse the same listener object in different VMs. Figure 14 shows the workflow of the iFogSim simulation. According to this experiment, we have one DC with four VMs to check the time of the user on cloud resources. This experiment is to check the progress of VMs that are used by cloud users.

6.5. CloudReports

The simulation set in this experiment consists of one DC and two hosts to demonstrate the capabilities of CloudReports (physical machines). Each host has a single CPU core (capable of 16,000 MIPS computing), 8 GB of RAM, and 12 TB of storage. In this scenario, a VM with 500 and 1000 MIPS compute volumes is running on the host. These VMs have 512 MB of RAM and 10 GB of storage for implementation. To begin this experiment in CloudReports, the DC must first change the host configuration and then add user privileges, as illustrated in Figure 15 and Figure 16. When the simulation is complete, CloudReports generates the results and displays them as HTML reports. Among them are charts for resource utilization, ET, power consumption, and overall resource utilization of each DC.
Figure 17 shows an analysis of the number of successful cloudlets executed in a VM with the current DC, and Figure 18 shows the total time in seconds for each successful cloudlet run in the current DC, both of which are generated by CloudReports. This experiment is based on data from cloud users related to the uploading time of cloud data into the cloud service. Based on the results, cloud users can upload the data in the fastest way because they have 16,000 MIPS of processing capacity that is based on a single-threshold allocation policy that is fast for the cloud users.

6.6. Result Analysis

Results for simulators with parameters are listed in Table 8, Table 9 and Table 10. Table 8 shows the results of the cloud simulation models discussed in this research. CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports are based on HPE, VM PE, and cloudlet PE parameters. Table 9 displays the findings of the cloud simulation models. CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports are based on UB parameters for Avg, Min, and Max with closet DC SB policies. Table 10 shows the results of CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports based on various ET analysis parameters for a customer cloudlet.

6.7. Performance Comparison of Simulators (Parameters)

The results received in Section 6 come from a comparison of five simulator metrics (HPEs; VM PEs; cloudlet PEs; UB Avg, Min, and Max; and cloudlet ID Start Time, Finish Time, Average Start, and Average Finish) for each of the five simulators (CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReport).
Figure 19 shows HPEs, VM PEs, and cloudlet PEs based on CloudSim, CloudSim Plus, and iFogSim. As it is shown, both CloudSim and CloudSimPlus are equal on HPEs, while the VM PEs and cloudlet PEs show a lower output value. Furthermore, iFogSim, compared to the other two simulators, is equal on VM PEs and cloudlet PEs, while the HPEs showed a higher output value than them.
Figure 20 shows a UB of 1 to 6 based on CloudAnalyst, as it shows UB 3 to UB 6 are different on Avg, Min, and Max, while UB 4 Min and UB 5 Min are almost equal. Furthermore, UB 1 Avg, Min, and Max and UB 2 Avg, Min, and Max are almost equal, while UB 3 Avg, Min, and Max and UB 6 Avg, Min, and Max showed a higher output value than them.
Figure 21 shows the cloudlet parameters: Start Time, Finish Time, Avg Start, and Avg Finish in seconds based on CloudReports. As it is shown, Cloudlet 0–50 start and finish times are equal, and Cloudlet 50–55 start, finish, Avg start, and Avg finish times are different, while the Cloudlet 55 finish time is equal to Cloudlet 0–50. Furthermore, Cloudlet 55 start, Avg start, and Avg finish times have different output values than those of Cloudlet 0–50.

7. Discussion

The results of the CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports simulators are discussed in this section. Cloud simulation parameters such as host, VM, cloudlet, UB, and ET, and comparison of simulation tools vs. non-simulation tools are also discussed. CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports enable different users to use a DC by employing various cloud-based simulations. Figure 7 shows the simulation capabilities of CloudSim while representing the graphical analysis of HPEs, VMPEs, and cloudlet PEs based on CloudSim. Figure 8 is based on CloudSim Plus. The CE is composed of six geographically dispersed UBs with DC 1 to 3, and three 5, 5, and 50 VM DCs, as shown in Figure 10. Based on the SB strategy, which picks the nearest DC in consideration of the lowest network delay, user-base traffic is routed to the DC. The simulation is 60 minutes long. After the simulation is done, CloudAnalyst displays the regions with time response data, as expressed in Figure 11. Analysis of the results can also be performed by exporting an a.pdf file, which includes comprehensive tables and charts. Figure 12 and Figure 13 are descriptive snapshots for each DC of a request’s servicing time and cost information, respectively. Figure 14 is based on iFogSim. Figure 15 and Figure 16 show the simulation environments of CloudReports, and Figure 17 and Figure 18 represent the number of cloudlets that the actual DC user has successfully performed using CloudReports and show the total time in seconds for each cloudlet to be successfully executed in the current DC.

7.1. Prominent Cloud Simulators

Q1: What is the motivation behind the simulator’s development?
Response: These simulators are modern and up-to-date, complete, and well-reported. Furthermore, they are cost-effective, provide results with graphical representation, and display effective performance.
Q2: What are the features or characteristics of cloud simulators?
Response: These are easy to use and expand, allowing users to model, simulate, and evaluate CC infrastructure and application services.
Q3: What is the scope of extension or application of the targeted cloud simulators?
Response: These simulators are used in PaaS, IaaS, and SAAS.

7.2. Research Questions

Q1: What parameters would be used for the comparison of simulation tools vs. non-simulation tools?
Response: We have used different parameters for the comparison of simulation tools vs. non-simulation tools such as host, VM, cloudlet, UB, and ET.
Q2: What metrics are covered in the cloud simulators that are targeted?
Response: We have covered different metrics for the cloud simulators, such as host, VMs, cloudlet, UBR, and ET.
Q3: What are the energy-consuming parameters found in the selected cloud simulators?
Response: We have found energy-consuming parameters such as workload, cluster, server host, VM, network, storage, CPU, and memory.

7.3. Simulation Setup of Cloud Simulators

We used CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports simulators for analysis to help researchers and readers in the real-time cloud environment. Table 11 shows the hardware specifications for an experiment.

8. Recommendations

Access to genuine cloud infrastructure is required for systems administrators, cloud experts, and even researchers to conduct real-time experiments and put new algorithms and approaches into practice. Before going live, it is critical to evaluate performance and thoroughly identify any potential security concerns. Modeling and simulation tools have saved us from such difficulties. A CC simulator is required to view an implementation scenario in real-time. Many CC simulators are being developed to assist academics, systems administrators, cloud specialists, and network administrators in assessing the real-time performance of CC environments. Since free and open-source simulators provide an environment for DL and experimentation, the trend is to focus on these simulators for all types of complicated and real-world CC problems.

9. Conclusions and Future Work

The ever-growing popularity and difficulty of computing systems make simulation software an essential option for creating, installing, controlling, and evaluating a system’s efficiency. Simulators of CC have been used in the evaluation of trade-offs between cost and reliability in PAYG settings. The analysis, therefore, aimed to present a collection of simulation resources appropriate for cloud architecture. A wide range of tools, namely CloudSim, CloudSim Plus, CloudAnalyst, iFogSim, and CloudReports, are addressed. Extensive analysis of cloud simulators that consider multiple hosts and VMs as different workloads and system configurations has been provided. The results of the performance comparison of simulator parameters are shown in Section 6.
In future work, the cloud simulators can be extended to develop innovative and advanced integrating technologies such as ML and AI to respond to the circumstances dynamically. Another area of research lies in expanding the cloud simulation tools’ functionality to fit ML algorithms, such as NB, RF, SVM, etc., in the simulation tools. These ML algorithms can be integrated with FTCloudSim, Dynamic CloudSim, etc. for cloud service reliability enhancement simulations. The enhancement of the simulation module graphics also has potential.

List of Abbreviations

Table 12 shows the list of abbreviations that are used throughout the whole manuscript. Details are mentioned below.

Author Contributions

Conceptualization, M.A.S.; Methodology, M.A.S. and M.M.A.; Validation, M.A.S. and M.M.A.; Investigation, M.A.S. and M.M.S.; Resources, M.A.S. and M.M.A.; Writing—original draft preparation, M.A.S. and M.M.A.; writing—review and editing, M.A.S., M.M.A. and M.M.S.; Visualization, M.A.S.; Supervision, M.A.S. and M.M.A. All authors have read and agreed to the published version of the manuscript.

Funding

The research will be funded by the Multimedia University, Department of Information Technology, Persiaran Multimedia, 63100, Cyberjaya, Selangor, Malaysia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would also like to thank their mentors who have always helped and guided this research to make it possible.

Conflicts of Interest

The authors declare that there are no conflict of interest.

References

  1. Bahwaireth, K.; Tawalbeh, L.; Benkhelifa, E.; Jararweh, Y.; Tawalbeh, M.A. Experimental comparison of simulation tools for efficient cloud and mobile cloud computing applications. J. Inf. Secur. 2016, 10, 15. [Google Scholar] [CrossRef] [Green Version]
  2. Mishra, S.K.; Sahoo, B.; Parida, P.P. Load balancing in cloud computing: A big picture. J. King Saud Univ. Comput. Inf. Sci. 2020, 32, 149–158. [Google Scholar] [CrossRef]
  3. Rashid, A.; Chaturvedi, A. Cloud Computing Characteristics and Services A Brief Review. Int. J. Comput. Sci. Eng. 2019, 7, 421–426. [Google Scholar] [CrossRef]
  4. Prajapati, P.; Sariya, A.K. A Review: Methods of Load Balancing on Cloud Computing. Int. J. Comput. Sci. Inf. Technol. 2019, 6, 8. [Google Scholar]
  5. Haris, M.; Khan, R.Z. A Systematic Review on Load Balancing Tools and Techniques in Cloud Computing. In Inventive Systems and Control; Suma, V., Baig, Z., Shanmugam, S.K., Lorenz, P., Eds.; Springer Nature: Singapore, 2022; pp. 503–521. [Google Scholar]
  6. Shahid, M.A.; Alam, M.M.; Su’ud, M.M. Performance Evaluation of Load-Balancing Algorithms with Different Service Broker Policies for Cloud Computing. Appl. Sci. 2023, 13, 1586. [Google Scholar] [CrossRef]
  7. Naik, N.; Mehta, M.A. Comprehensive and Comparative Study of Cloud Simulators. In Proceedings of the 2018 IEEE Punecon, Pune, India, 30 November–2 December 2018; pp. 1–7. [Google Scholar]
  8. Zaidi, T. Analysis of Energy Consumption on IaaS Cloud Using Simulation Tool. SSRN J. 2020, 1, 1–12. [Google Scholar] [CrossRef]
  9. Tissir, N.; ElKafhali, S.; Aboutabit, N. How Much Your Cloud Management Platform Is Secure? OpenStack Use Case. In Innovations in Smart Cities Applications Volume 4; Ahmed, M.B., Kara, İ.R., Santos, D., Sergeyeva, O., Boudhir, A.A., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 183, pp. 1117–1129. [Google Scholar]
  10. Bambrik, I. A Survey on Cloud Computing Simulation and Modeling. SN Comput. Sci. 2020, 1, 249. [Google Scholar] [CrossRef]
  11. Mansouri, N.; Ghafari, R.; Zade, B.M.H. Cloud computing simulators: A comprehensive review. Simul. Model. Pract. Theory 2020, 104, 102144. [Google Scholar] [CrossRef]
  12. Sundas, A.; Panda, S.N. An Introduction of CloudSim Simulation tool for Modelling and Scheduling. In Proceedings of the 2020 International Conference on Emerging Smart Computing and Informatics (ESCI), Pune, India, 12–14 March 2020; pp. 263–268. [Google Scholar]
  13. Puhan, S.; Panda, D.; Mishra, B.K. Energy Efficiency for Cloud Computing Applications: A Survey on the Recent Trends and Future Scopes. In Proceedings of the 2020 International Conference on Computer Science, Engineering and Applications (ICCSEA), Gunupur, India, 13–14 March 2020; pp. 1–6. [Google Scholar]
  14. Hassaan, M. A Comparative Study between Cloud Energy Consumption Measuring Simulators. Int. J. Educ. Manag. Eng. 2020, 10, 20–27. [Google Scholar] [CrossRef]
  15. Ismail, A. Energy-driven cloud simulation: Existing surveys, simulation supports, impacts and challenges. Clust. Comput. 2020, 23, 3039–3055. [Google Scholar] [CrossRef]
  16. Silva Filho, M.C.; Oliveira, R.L.; Monteiro, C.C.; Inacio, P.R.M.; Freire, M.M. CloudSim Plus: A cloud computing simulation framework pursuing software engineering principles for improved modularity, extensibility and correctness. In Proceedings of the 2017 IFIP/IEEE Symposium on Integrated Network and Service Management (IM), Lisbon, Portugal, 8–12 May 2017; pp. 400–406. [Google Scholar]
  17. Markus, A.; Kertesz, A. A survey and taxonomy of simulation environments modelling fog computing. Simul. Model. Pract. Theory 2020, 101, 102042. [Google Scholar] [CrossRef]
  18. Anusooya, G.; Md, A.Q.; Jackson, C.; Prassanna, J. A review on effective utilization of computational resources using cloudsim. J. Crit. Rev. 2020, 7, 1–7. [Google Scholar]
  19. Nandhini, J.M.; Gnanasekaran, T. An Assessment Survey of Cloud Simulators for Fault Identification. In Proceedings of the 2019 3rd International Conference on Computing and Communications Technologies (ICCCT), Chennai, India, 21–22 February 2019; pp. 311–315. [Google Scholar]
  20. Arseniev, D.G.; Overmeyer, L.; Kälviäinen, H.; Katalinić, B. Cyber-Physical Systems and Control; Springer International Publishing: Cham, Switzerland, 2019; Volume 95. [Google Scholar]
  21. Bala, M.I.; Chishti, M.A. Offloading in Cloud and Fog Hybrid Infrastructure Using iFogSim. In Proceedings of the 2020 10th International Conference on Cloud Computing, Data Science and Engineering (Confluence), Noida, India, 29–31 January 2020; pp. 421–426. [Google Scholar]
  22. Seo, D.; Shahhosseini, S.; Mehrabadi, M.A.; Donyanavard, B.; Lim, S.S.; Rahmani, A.M.; Dutt, N. Dynamic iFogSim: A Framework for Full-Stack Simulation of Dynamic Resource Management in IoT Systems. In Proceedings of the 2020 International Conference on Omni-layer Intelligent Systems (COINS), Barcelona, Spain, 31 August–2 September 2020; pp. 1–6. [Google Scholar]
  23. Fakhfakh, F.; Kacem, H.H.; Kacem, A.H. Simulation tools for cloud computing: A survey and comparative study. In Proceedings of the 2017 IEEE/ACIS 16th International Conference on Computer and Information Science (ICIS), Wuhan, China, 24–26 May 2017; pp. 221–226. [Google Scholar]
  24. Motlhabane, N.; Gasela, N.; Esiefarienrhe, M. Comparative Analysis of Cloud Computing Simulators. In Proceedings of the 2018 International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, USA, 12–14 December 2018; pp. 1309–1316. [Google Scholar]
  25. Byrne, J.; Svorobej, S.; Giannoutakis, K.M.; Tzovaras, D.; Byrne, P.J.; Östberg, P.O.; Gourinovitch, A.; Lynn, T. A Review of Cloud Computing Simulation Platforms and Related Environments. In Proceedings of the 7th International Conference on Cloud Computing and Services Science, Porto, Portugal, 24–26 April 2017; pp. 679–691. [Google Scholar]
  26. Khalil, K.M.; Abdel-Aziz, M.; Nazmy, T.T.; Salem, A.-B.M. Cloud simulators—An evaluation study. Int. J. Inf. Model. Anal. 2017, 6, 23. [Google Scholar]
  27. Dogra, S.; Singh, A.J. Comparison of Cloud Simulators for effective modeling of Cloud applications. Simul. Model. Pract. Theory 2020, 99, 1–13. [Google Scholar]
  28. Perez Abreu, D.; Velasquez, K.; Curado, M.; Monteiro, E. A comparative analysis of simulators for the Cloud to Fog continuum. Simul. Model. Pract. Theory 2020, 101, 102029. [Google Scholar] [CrossRef]
  29. Gupta, H.; Vahid Dastjerdi, A.; Ghosh, S.K.; Buyya, R. iFogSim: A toolkit for modeling and simulation of resource management techniques in the Internet of Things, Edge and Fog computing environments: iFogSim: A toolkit for modeling and simulation of internet of things. Softw. Pract. Exper. 2017, 47, 1275–1296. [Google Scholar] [CrossRef] [Green Version]
  30. Makaratzis, A.T.; Giannoutakis, K.M.; Tzovaras, D. Energy Modeling in Cloud Simulation Frameworks. Future Gener. Comput. Syst. 2018, 79, 715–725. [Google Scholar] [CrossRef]
  31. Jena, S.R.; Shanmugam, R.; Saini, K.; Kumar, S. Cloud Computing Tools: Inside Views and Analysis. Procedia Comput. Sci. 2020, 173, 382–391. [Google Scholar] [CrossRef]
  32. Esparcia, J.A.P.; Singh, M. Comprehensive study of multi-resource cloud simulation tools. Int. J. Adv. Appl. Sci. 2017, 4, 29–38. [Google Scholar] [CrossRef]
  33. Ashalatha, R.; Agarkhed, J.; Patil, S. Analysis of Simulation Tools in Cloud Computing. In Proceedings of the 2016 International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET), Chennai, India, 23–25 March 2016; pp. 748–751. [Google Scholar]
  34. Beena, T.L.A.; Lawanya, J.J. Simulators for Cloud Computing—A Survey. IEEE Access 2018, 7, 1–7. [Google Scholar]
  35. Suryateja, P.S. A Comparative Analysis of Cloud Simulators. Int. J. Mod. Educ. Comput. Sci. 2016, 8, 64–71. [Google Scholar] [CrossRef]
  36. Khurana, R.; Bawa, R.K. Quality based cloud simulators: State-of-the-art road ahead. In Proceedings of the 2016 Fourth International Conference on Parallel, Distributed and Grid Computing (PDGC), Solan, India, 22–24 December 2016; pp. 101–106. [Google Scholar]
  37. Khan, A.A.; Aleem, M.; Sajjad, A. Energy-Aware Cloud Computing Simulators: A State of the Art Survey. Int. J. Mod. Educ. Comput. Sci. 2018, 6, 15–20. [Google Scholar]
  38. Sumitha, J. Study of Simulation Tools in Cloud Computing Environment. J. Indep. Stud. Res. Comput. 2018, 7, 284–289. [Google Scholar]
  39. Shakir, M.S.; Razzaque, E.A. Performance Comparison of Load Balancing Algorithms using Cloud Analyst in Cloud Computing. IEEE Access 2017, 5, 1–5. [Google Scholar]
  40. Vashistha, A.; Sholliya, S. Comparative study of open source cloud simulation tools. Int. J. Eng. Res. 2017, 4, 5. [Google Scholar]
  41. Kumar, K.A. A Study on Simulation Tools in Cloud Computing. J. Glob. Res. Comput. Sci. 2017, 7, 1–7. [Google Scholar]
  42. Malhotra, M. A Study and Analysis on Simulators of Cloud Computing Paradigm. Int. J. Adv. Trends Appl. 2017, 4, 3. [Google Scholar]
  43. Maarouf, A.; Marzouk, A.; Haqiq, A. Comparative Study of Simulators for Cloud Computing. In Proceedings of the 2015 International Conference on Cloud Technologies and Applications (CloudTech), Marrakech, Morocco, 2–4 June 2015; pp. 1–8. [Google Scholar]
  44. Pagare, J.D.; Koli, D.N.A. Design and simulate cloud computing environment using cloudsim. Int. J. Recent Technol. Eng. 2015, 6, 8. [Google Scholar]
  45. Gupta, K.; Beri, R. Cloud Computing: A Survey on Cloud Simulation Tools. J. Glob. Res. Comput. Sci. 2016, 2, 430–434. [Google Scholar]
  46. Singh, D.A.A.G.; Priyadharshini, R.; Leavline, E.J. Analysis of Cloud Environment Using CloudSim. In Artificial Intelligence and Evolutionary Computations in Engineering Systems; Dash, S.S., Naidu, P.C.B., Bayindir, R., Das, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2018; pp. 325–333. [Google Scholar]
  47. Andrade, E.; Nogueira, B. Performability Evaluation of a Cloud-Based Disaster Recovery Solution for IT Environments. J. Grid Comput. 2019, 17, 603–621. [Google Scholar] [CrossRef]
  48. Ahmad, M.O.; Khan, R.Z. Cloud Computing Modeling and Simulation Using CloudSim Environment. Int. J. Recent Technol. Eng. 2019, 8, 5439–5445. [Google Scholar] [CrossRef]
  49. Bhatia, M.S.; Sharma, M. A Critical Review and Analysis of Cloud Computing Simulators. Int. J. Latest Trends Eng. Technol. 2016, 1, 29–36. [Google Scholar]
  50. Shahid, M.A.; Islam, N.; Alam, M.M.; Su’ud, M.M.; Musa, S. A Comprehensive Study of Load Balancing Approaches in the Cloud Computing Environment and a Novel Fault Tolerance Approach. IEEE Access 2020, 8, 130500–130526. [Google Scholar] [CrossRef]
  51. Shahid, M.A.; Islam, N.; Alam, M.M.; Mazliham, M.S.; Musa, S. Towards Resilient Method: An exhaustive survey of fault tolerance methods in the cloud computing environment. Comput. Sci. Rev. 2021, 40, 100398. [Google Scholar] [CrossRef]
  52. Núñez, A.; Cañizares, P.C.; de Lara, J. CloudExpert: An intelligent system for selecting cloud system simulators. Expert Syst. Appl. 2022, 187, 115955. [Google Scholar] [CrossRef]
  53. Singh, H.; Tyagi, S.; Kumar, P. Comparative Analysis of Various Simulation Tools Used in a Cloud Environment for Task-Resource Mapping. In Proceedings of the International Conference on Paradigms of Computing, Communication and Data Sciences, Kurukshetra, India, 7–9 May 2021; Springer Nature: Singapore, 2021; pp. 419–430. [Google Scholar]
  54. Saadi, Y.; El Kafhali, S. Energy-efficient strategy for virtual machine consolidation in cloud environment. Soft Comput. 2020, 24, 14845–14859. [Google Scholar] [CrossRef]
  55. Lago, D.G.; da Silva, R.A.C.; Madeira, E.R.M.; da Fonseca, N.L.S.; Medhi, D. SinergyCloud: A simulator for evaluation of energy consumption in data centers and hybrid clouds. Simul. Model. Pract. Theory 2021, 110, 102329. [Google Scholar] [CrossRef]
  56. Satapathy, S.; Bhateja, V.; Mohanty, J.R.; Udgata, S.K. Smart Intelligent Computing and Applications: Proceedings of the Third International Conference on Smart Computing and Informatics, Volume 1; Springer: Singapore, 2020; Volume 159. [Google Scholar]
  57. Giannoutakis, K.M.; Filelis-Papadopoulos, C.K.; Gravvanis, G.A.; Tzovaras, D. Evaluation of Self-Organizing and Self-Managing Heterogeneous High Performance Computing Clouds through Discrete-Time Simulation. Concurr. Comput. Pract. Exp. 2021, 33, e6326. [Google Scholar] [CrossRef]
Figure 1. Categories of cloud simulators [10].
Figure 1. Categories of cloud simulators [10].
Applsci 13 08785 g001
Figure 2. The architecture of CloudSim [34,42].
Figure 2. The architecture of CloudSim [34,42].
Applsci 13 08785 g002
Figure 3. CloudSim Plus API package kit [16].
Figure 3. CloudSim Plus API package kit [16].
Applsci 13 08785 g003
Figure 4. CloudAnalyst architecture [25,36].
Figure 4. CloudAnalyst architecture [25,36].
Applsci 13 08785 g004
Figure 5. iFogSim architecture [22].
Figure 5. iFogSim architecture [22].
Applsci 13 08785 g005
Figure 6. CloudReports architecture [1].
Figure 6. CloudReports architecture [1].
Applsci 13 08785 g006
Figure 7. Framework for analysis and comparison.
Figure 7. Framework for analysis and comparison.
Applsci 13 08785 g007
Figure 8. Simulation results of CloudSim based on the power usage of hosts and VMs.
Figure 8. Simulation results of CloudSim based on the power usage of hosts and VMs.
Applsci 13 08785 g008
Figure 9. Simulation results of CloudSim Plus based on the round-robin VM allocation policy.
Figure 9. Simulation results of CloudSim Plus based on the round-robin VM allocation policy.
Applsci 13 08785 g009
Figure 10. Settings of simulation duration with closet DC SB policy.
Figure 10. Settings of simulation duration with closet DC SB policy.
Applsci 13 08785 g010
Figure 11. Simulation results of DC request servicing time [1].
Figure 11. Simulation results of DC request servicing time [1].
Applsci 13 08785 g011
Figure 12. Simulation results of DC request servicing time.
Figure 12. Simulation results of DC request servicing time.
Applsci 13 08785 g012
Figure 13. Simulation results of DC, VM, and data transfer costs.
Figure 13. Simulation results of DC, VM, and data transfer costs.
Applsci 13 08785 g013
Figure 14. Simulation results of iFogSim based on the DC with one host that places two VMs.
Figure 14. Simulation results of iFogSim based on the DC with one host that places two VMs.
Applsci 13 08785 g014
Figure 15. Simulation results of the DC based on the single threshold allocation policy.
Figure 15. Simulation results of the DC based on the single threshold allocation policy.
Applsci 13 08785 g015
Figure 16. Simulation results of customers based on VMs.
Figure 16. Simulation results of customers based on VMs.
Applsci 13 08785 g016
Figure 17. Simulation results of the number of successful cloudlets executed in a VM with the current DC.
Figure 17. Simulation results of the number of successful cloudlets executed in a VM with the current DC.
Applsci 13 08785 g017
Figure 18. Simulation results of ET for the customer’s cloudlet.
Figure 18. Simulation results of ET for the customer’s cloudlet.
Applsci 13 08785 g018
Figure 19. Performance comparison of the host, VM, and cloudlet PEs.
Figure 19. Performance comparison of the host, VM, and cloudlet PEs.
Applsci 13 08785 g019
Figure 20. Performance comparison of six UBs based on Avg, Min, and Max values.
Figure 20. Performance comparison of six UBs based on Avg, Min, and Max values.
Applsci 13 08785 g020
Figure 21. Performance comparison of cloudlet ID based on start, finish, Avg start, and Avg finish times in seconds.
Figure 21. Performance comparison of cloudlet ID based on start, finish, Avg start, and Avg finish times in seconds.
Applsci 13 08785 g021
Table 1. A list of key questions.
Table 1. A list of key questions.
IDSubtopicsQuestions
1Prominent Cloud SimulatorsWhat is the motivation behind the simulator’s development?
What are the features or characteristics of cloud simulators?
What is the scope of extension or application of the targeted cloud simulators?
2DiscussionWhat parameters would be used for the comparison of simulation tools vs. non-simulation tools?
3Energy-Aware ProvisioningWhat metrics are covered in the cloud simulators that are targeted?
What are the energy-consuming parameters found in the selected cloud simulators?
Table 2. Advantages and disadvantages of cloud simulators [16,16,25,26,27,28].
Table 2. Advantages and disadvantages of cloud simulators [16,16,25,26,27,28].
SimulatorsAdvantagesDisadvantages
CloudSim
  • This is a basic simulator and is perfect for batch task simulation [27].
  • Easy to use and quickly expanded toolkit for the simulation [27].
  • Provides basic modeling, simulation, and testing [27].
  • Streamed activities cannot be assisted (i.e., by continuous computation) [27].
  • Not supported by PaaS and SaaS [26].
  • There is no graphical output [16].
CloudSim Plus
  • It is a modern simulation framework that is up-to-date, complete, and well-reported [16].
  • It is simple to use and extend, allowing CC infrastructures and application services to be modeled, simulated, and evaluated [16].
Does not provide graphical output [16].
CloudAnalyst
  • It assists in the simulation and study of CC systems on a wide-scale [27].
  • Effective performance is generated very easily and rapidly [27].
  • Provides graphical output [27].
  • No migration policy was provided [26].
  • Does not support PaaS and SaaS [26].
iFogSim
  • Low cost [29].
  • Provides graphical output [28].
  • Provides stream processing [28].
  • Provides data streaming [28].
  • Does not provide distributed architecture [25].
  • Does not provide web API availability [25].
CloudReports
  • Provides graphical output [27].
  • The environment is user-friendly and offers more customization and repeatability based on various parameters [27].
  • Does not support PaaS [26].
  • No migration policy was provided [26,30].
Table 3. Selection of cloud simulators: a simulation-based assessment of the specifications [26,29,31,32,33].
Table 3. Selection of cloud simulators: a simulation-based assessment of the specifications [26,29,31,32,33].
SimulatorKind of
Simulator
Side-by-Side AnalysisTime of SimulationAvailabilityHdwe/Soft
Associated
Programing LanguageFundamental
Policy
Analysis
Repetition
GUIUser
Preferences
Dynamic Run
CloudSimevent-basedSecondsSoftwareJavaSimJava
CloudSim Plusevent-basedSecondsSoftwareJavaSimJava
CloudAnalystevent-basedSecondsSoftwareJavaCloudSim/
SimJava
iFogSimevent-basedSecondsSoftwareJavaCloudSim/
SimJava
CloudReportsevent-basedSecondsSoftwareJavaCloudSim
Table 4. The following is a list of the energy-consuming parameters found in the selected cloud simulators [11,15,26,31].
Table 4. The following is a list of the energy-consuming parameters found in the selected cloud simulators [11,15,26,31].
SimulatorWorkloadClusterServer/HostVirtual MachineNetworkStorageCPUMemory
CloudSim
CloudSim Plus
CloudAnalyst
iFogSim
CloudReports
Table 7. A summary of various cloud simulator architectures [1,16,22,35,49].
Table 7. A summary of various cloud simulator architectures [1,16,22,35,49].
RefNameYearSimulatorConcise Description of Architecture
[1]Khadijah Bahwairethet al.2016CloudReportsAnother instrument that is based on CloudSim is CloudReport. CloudReport, however, improves many of CloudSim’s features, such as a friendly GUI, running multiple simulations, and producing better simulations [1].
[16]Manoel C. Silva Filho et al.2017CloudSim Plus
  • CloudSim Plus is a Java 8 simulation platform that enables various CC services, ranging from IaaS to SaaS layers, to be modeled and simulated [16].
  • It allows simulation scenarios to be introduced for experimentation, evaluation, and validation of algorithms for various objectives [16].
[22]Dongjoo Seo et al.2020iFogSim
  • Several components are included in the proposed architecture: (1) an expanded iFogSim simulator; (2) a Sniper Multi-Core simulator; (3) a Cross-layer system Controller; and (4) a Co-simulation Environment module [22].
  • For the specified system configuration, the architecture reports the application’s ET and energy consumption [22].
[35]Pericherla S Suryateja2016CloudAnalyst
  • CloudAnalyst provides an excellent GUI for configuring any geographically distributed system, which includes application workloads, the geographic location of users generating traffic, the DC location, the number of users per DC, and the amount of resources per DC. During the simulation time, CloudAnalyst can create results as a chart or table that highlight the massive quantity of users and system statistics [35].
[49]Mandeep S. Bhatia and Manmohan Sharma2016CloudSim
  • The user code that handles the problems related to simulation requirements and scheduling policies is the top layer. This layer thus implements the specifications of the user, such as considering the number of machines and assigning hosts to VMs [49].
  • The middle layer provides support for cloud-based environment development and simulation, which includes VM interfaces, storage, and bandwidth [49].
  • The lowest layer is referred to as the simulation engine since it contains the core jobs required for high-level simulation settings [49].
Table 8. Processing element comparison of cloud simulators.
Table 8. Processing element comparison of cloud simulators.
SimulatorsParameters
Host Processing ElementsVM Processing
Elements
Cloudlet Processing Elements
CloudSim842
CloudSim Plus822
CloudAnalyst613
iFogSim411
CloudReports721
Table 9. UB parameter comparison of cloud simulators.
Table 9. UB parameter comparison of cloud simulators.
SimulatorsParameters
Avg.Min.Max.
CloudSim55.86544.2
CloudSim Plus5258.844.8
CloudAnalyst74.3104.551.8
iFogSim355.4425.8257.6
CloudReports313.6363.3268.3
Table 10. Cloudlet parameter comparison of cloud simulators.
Table 10. Cloudlet parameter comparison of cloud simulators.
SimulatorsCloudlet ID Parameters
Start Time (s)End Time (s)AVG Start (s)AVG Finish (s)
CloudSim00500500
CloudSim Plus1100100500500
CloudAnalyst11000500500
iFogSim0100500500
CloudReports0100500500
Table 11. Hardware specifications for the experiment.
Table 11. Hardware specifications for the experiment.
System
ProcessorIntel (R) Core (TM) i7 2620M CPU 2.7 GHz
RAM8.00 GB
Windows8.1 Platform
IDE-1Eclipse IDE for Java Developer Release 2021-09 (4.21.0)
Table 12. List of abbreviations.
Table 12. List of abbreviations.
AIArtificial Intelligence
AppsApplications
APIApplication Programming Interface
AWSAmazon Web Services
AVGAverage
CCCloud Computing
CSPCloud Service Provider
CPUCentral Processing Unit
DLDeep Learning
CECloud Environment
DCData left
DCsData lefts
CDCCloset Data left
ETExecution Time
FTFault Tolerance
GUIGraphical User Interface
HPEHost Processing Element
HPEsHost Processing Elements
IaaSInfrastructure as a Service
IoTInternet of Things
LBLoad Balancing
MLMachine Learning
MinMinimum
MaxMaximum
NASANational Aeronautics and Space Administration
NBNaïve Bayes
PaaSPlatform as a Service
PAYGPay-as-you-go
PEProcessing Element
PEsProcessing Elements
RFRandom Forest
SVMSupport Vector Machine
SBService Broker
SaaSSoftware as a Service
SBPService Broker Policy
SBPsService Broker Policies
UBRUser Base Response
UBUser Base
UBsUser Bases
VMVirtual Machine
VMsVirtual Machines
VMPEsVirtual Machine Processing Elements
QoSQuality of Service
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shahid, M.A.; Alam, M.M.; Su’ud, M.M. A Systematic Parameter Analysis of Cloud Simulation Tools in Cloud Computing Environments. Appl. Sci. 2023, 13, 8785. https://doi.org/10.3390/app13158785

AMA Style

Shahid MA, Alam MM, Su’ud MM. A Systematic Parameter Analysis of Cloud Simulation Tools in Cloud Computing Environments. Applied Sciences. 2023; 13(15):8785. https://doi.org/10.3390/app13158785

Chicago/Turabian Style

Shahid, Muhammad Asim, Muhammad Mansoor Alam, and Mazliham Mohd Su’ud. 2023. "A Systematic Parameter Analysis of Cloud Simulation Tools in Cloud Computing Environments" Applied Sciences 13, no. 15: 8785. https://doi.org/10.3390/app13158785

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop