Next Article in Journal
Performance Analysis of OPC UA for Industrial Interoperability towards Industry 4.0
Previous Article in Journal
Optimizing Trajectory and Dynamic Data Offloading Using a UAV Access Platform
Previous Article in Special Issue
Performance Evaluation of Federated Learning for Residential Energy Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Defining and Assessing Quality in IoT Environments: A Survey

by
Aggeliki Sgora
1,* and
Periklis Chatzimisios
2,3
1
Department of Digital Media and Communication, Ionian University, 28100 Argostoli, Greece
2
Department of Information and Electronic Systems Engineering, International Hellenic University of Greece (IHU), Alexander Campus, 574 00 Sindos, Greece
3
Department of Electrical and Computer Engineering, The University of New Mexico, Albuquerque, NM 87131-0001, USA
*
Author to whom correspondence should be addressed.
IoT 2022, 3(4), 493-506; https://doi.org/10.3390/iot3040026
Submission received: 4 July 2022 / Revised: 22 November 2022 / Accepted: 23 November 2022 / Published: 7 December 2022
(This article belongs to the Special Issue Advanced Quality of Service Approaches in Edge Computing)

Abstract

:
With the proliferation of multimedia services, Quality of Experience (QoE) has gained a lot of attention. QoE ties together the users’ needs and expectations to multimedia application and network performance. However, in various Internet of Things (IoT) applications such as healthcare, surveillance systems, traffic monitoring, etc., human feedback can be limited or infeasible. Moreover, for immersive augmented and virtual reality, as well as other mulsemedia applications, the evaluation in terms of quality cannot only focus on the sight and hearing senses. Therefore, the traditional QoE definition and approaches for evaluating multimedia services might not be suitable for the IoT paradigm, and more quality metrics are required in order to evaluate the quality in IoT. In this paper, we review existing quality definitions, quality influence factors (IFs) and assessment approaches for IoT. This paper also introduces challenges in the area of quality assessment for the IoT paradigm.

1. Introduction

Quality of Service (QoS), according to the International Telecommunications Union (ITU), is “the totality of characteristics of a telecommunications service that bear on its ability to satisfy stated and implied needs of the user of the service” [1]. It comprises both network-related performance (e.g., bit error rate, latency) and non-network related performance (e.g., service provisioning time, different tariffs) [2]. Thus, in order to satisfy the users’ needs, for more than 30 years the telecommunications industry, as well as academia, investigated several mechanisms in order to guarantee QoS in the provided telecommunication services.
However, with the exponential growth of the video-based services, the telecom operators realized that catering the quality expectation of the end users in multimedia services is the most important parameter [3]. Humans are considered quality meters, and their expectations, perceptions, and needs with respect to a particular product, service, or application carry great value [4,5].
The user experience of multimedia applications is inevitably bounded up with the notion of Quality of Experience (QoE) [6]. Lagjhari et al. described QoE as “the blueprint of all human quality needs and expectations” [4].
However, with the introduction of the Internet of Things (IoT), traditional terms and approaches used for defined or evaluating services may not be suitable/sufficient for the IoT context [7], where consumers may no longer be users but machines. Moreover, the QoE requirements in such a heterogeneous environment can vary with respect to the considered IoT application domain; even QoE requirements among IoT applications of the same IoT domain may vary [8]. Furthermore, due to the fact that, in IoT, decisions are taken based on data infusion from multiple sensors in case of failures, the effects are often multidimensional [9].
Currently, there is no standardization or set of best practices as to how the subjective tests can be conducted and even if there was, it would be practically infeasible to carry out subjective tests for every existing as well as new applications [7]. Moreover, existing subjective methodologies do not consider QoE influence factors (IFs) of the IoT environment, such as the usefulness of the application [7].
Thus, in such heterogeneous environments, existing quality related issues that were initially tailored for humans, such as QoE definitions, evaluation approaches and provision mechanisms, should be re-examined in order to check their validity in a machine-to machine environment. In addition, in cases where human interaction is not required, the traditional definition of QoE is not valid, and new metrics to evaluate the quality in IoT environments are required.
In the literature, several survey papers may be found that focus on IoT, e.g., [10,11]) or QoE, e.g., [12,13,14]). However, there are few survey papers that examine QoE in IoT environments.
Fizza et al. [15] reviewed existing QoE definitions and QoE models for autonomic IoT. However, the authors suggested only one definition for QoE in IoT. In addition, regarding the QoE modeling, the authors provided limited information concerning the role of data in the QoE evaluation of the IoT applications. Moreover, the authors in [7] focused on the QoE IFs and presented a QoE taxononomy for IoT, while Bures et al. [16] consolidated the IoT quality characteristics into a unified view. A survey concerning QoE evaluation for autonomic IoT applications can be found in [17].
The contributions of the current paper are as follows:
  • Review existing definitions of QoE that are suitable for IoT environments, since nowadays new terms have been introduced to define and evaluate the quality of IoT applications.
  • Identify and categorize the quality IFs for IoT. More specifically, we have collected and classified IFs that may found into the literature and are necessary for the creation of a successful quality model for IoT.
  • Review existing quality assessment approaches for IoT applications.
The rest of this paper is organized as follows: Section 1 presents an overview of IoT and fog computing architectures. Section 2 reviews existing quality definitions that are suitable for IoT, while Section 3 overviews and categorizes quality indicators for IoT. Section 4 reviews existing QoE models and frameworks for IoT. Section 5 discusses challenges in the area of QoE assessment in the IoT context, while Section 6 concludes the paper.

2. Internet of Things (IoT) and Fog Computing

The term “Internet of Things” (IoT) was coined by Kevin Ashton in 1999 to describe the ability of network objects (connected to the Internet) to bring new services [18]. Since then, the IoT paradigm has gained a great momentum. More specifically, according to the statistics portal Statista (www.statista.com (accessed on 1 June 2022)), the number of IoT connected devices is expected to rise more than 75 billion in 2025 [19]. Figure 1 illustrates several fields of IoT applications including transportation, healthcare, home automation and smart cities [20]. However, these diverse IoT applications and devices from various manufacturers create such network heterogeneity that a unified and inter-operable standard is very difficult to achieve [10].
Currently, there is no global consensus on the architecture of IoT, thus, many different IoT architectures may be found in the literature [11]. The basic architecture has three layers [10]:
  • A things layer (also known as perception, device or sensor layer) that consists of the sensing hardware, and its main objective is to interconnect things in the IoT network.
  • A middle layer (also known as transport layer) that processes the received data from the things layer and determines the optimum data transmission path to the IoT servers.
  • An application layer (also known as the business layer) that provides information management, data mining, data analytics, decision-making services, as well as the required services to end-user or machines.
However, according to Kassab et al. [11] the superior model with respect to the elements is the architecture proposed by Yousefpour et al. (2018) [21], a fog computing architecture. Fog computing is a computing paradigm (introduced by Cisco) that deals with the requirements of time-sensitive IoT applications [22]. The idea is that instead of processing the sensor data on the cloud, to address this issue at the edge [23]. By doing so, the following advantages are accomplished [22]: i) applications are executed closer to end-user and IoT devices, (ii) performance metrics for real-time applications such as latency, response time, and cloud workload are enhanced, (iii) network scalability is increased, and (iv) device mobility is supported.
Figure 2 depicts the three layers of the fog computing architecture. The lowest layer consists of the IoT devices that produce massive data and potentially are heterogeneous, geographically distributed and have mobility features [24]. The fog computing layer is composed by the fog nodes, intelligent intermediate devices from different networking components [25] and retransmits the workload to the cloud servers at given time intervals.

3. Quality in a IoT Environment

In telecommunications, the most suitable metric to assess end-to-end quality is QoE. The most frequently used definition for QoE is the one given by the ITU [26], where it is defined as “The overall acceptability of an application or service, as perceived subjectively by the end-user”. Moreover, since many researchers pointed out that the inclusion of the term “acceptability” as the basis for a QoE definition is not ideal, during the Dagstuhl Seminar in 2009, the term acceptability was newly defined as “the outcome of a decision [yes/no] which is partially based on the Quality of Experience” [27]. However, even with this modification, the definition still follows a user-centric approach; thus, it does not reflect the machines’ perspective.
Another popular definition of QoE is the one described in the Qualinet White paper [28] in which QoE is defined as “the degree of delight or annoyance of the user of an application or service. It results from the fulfillment of his or her expectations with respect to the utility and/or enjoyment of the application or service in the light of the user’s personality and current state”. Raake and Eggger [27] extended the definition of the Qualinet White paper in order to also include the term system. Thus, according to the new definition QoE is “the degree of delight or annoyance of a person whose experiencing involves an application, service, or system. It results from the person’s evaluation of the fulfillment of his or her expectations and needs with respect to the utility and/or enjoyment in the light of the person’s context, personality and current state”.
The Qualinet’s definition according to Floris and Atzori [20] is valid for general multimedia applications/services and, thus, it can be used for cases where humans are the recipients of the content provided by multimedia IoT applications.
However, in the IoT context, where exist applications that do not require any human intervention, such as smart parking, connected vehicles, etc., the term QoE cannot be used to describe quality. To this end, several researchers introduced new terms to define quality in the IoT domain. Mivoski et al. [29] introduced the term Quality of IoT-experience (QoIoT), which aggregates the delivered quality of an IoT service from the perspective of both humans and machines within the context of autonomous vehicles. More specifically, the QoIoT metric compromises the traditional user-centric QoE metric and the Quality of Machine Experience (QoME), an objective metric that “measures the quality and performance of intelligent machines and their decisions”.
Karaadi et al. [30] defined the term “Quality of Things” (QoT) for multimedia communications in IoT to express the quality of fulfilling an IoT task in a Multimedia IoT (M-IoT) [31]. However, the authors do not provide any measurement methodology.
Rahman et al. [9] defined the term Quality of Systems (QoSys), an objective metric like QoE, that measures “the quality and performance of the Systems of Systems (SoS), and the decisions made by those”. Thus, the metric Q o E I O T is introduced in order to evaluate the quality in an IoT scenario from the perspective of both humans and machines.
Wang et al. [32] introduced the term quality of X (QoX), as a comprehensive evaluation metric that combines QoS, QoE, Quality of Data (QoD) and Quality of Information (QoI).
In this end, Fizza et al. [17] introduce the term Quality of autonomic IoT applications as “an aggregate quantitative value of various IoT quality metrics measured at each stage of the autonomic IoT application life cycle”.
Table 1 overviews and highlights the specific drawbacks of each definition. As can be seen, there is no definition that can generally express the end-to-end quality in IoT environments.

4. Key Quality Indicators for IoTs

As stated in [23], the first step in creating a successful quality model is to create a taxonomy of its influence factors (IFs). However, identifying these factors is not an easy task to accomplish.
As concerns QoE, existing approaches classify IFs into multiple dimensions. Stankiewicz et al. [33] identified that the factors that impact the QoE of multimedia services are QoS, Grade of Service (GoS), Quality of Resilience (QoR), as well as several orthogonal factors depicted in Figure 3.
Le Callet et al. [28] classify the QoE IFs into the following 3 categories [5]:
  • Human IFs that present any variant or invariant property or characteristic of a human user (e.g., motivation, gender, age, education, etc.);
  • System IFs that refer to properties and characteristics that determine the technically produced quality of an application or service (e.g., QoS, display size, resolution);
  • Context IFs that embrace any situational property to describe the user’s environment in terms of physical, temporal, social, economic, task, and technical characteristics (e.g., day of time, cost, etc.).
However, since in IoT the data acquired by devices (objects), as well as the information acquired and processed are important parameters, two more categories/dimensions may be found in the literature: the Quality of Data (QoD) that is used for data quality evaluation, and the Quality of Information (QoI) that is used for information quality evaluation. However, in several papers, the term QoI is used to determine the quality of information or data [34,35].
Table 2 overviews the most common QoD metrics, while Table 3 shows the most common QoI metrics.
Fizza et al. [17] considered the following IFs for the autonomic IoT applications: (1) the QoD; (2) the Quality of Device (QoDe); (3) the QoS; (4) the Quality of Context (QoC); (5) the QoI (6) Quality of Security and Privacy (QoSe & P); and (7) Quality of Actuation (QoA).
Rahman et al. [9] also considered the Quality of Cost (QoC) due to the fact that the machines use some resources in terms of computation, storage, or energy, and such consumptions should be optimized.
Ikeda et al. [36] considered two sets of metrics: physical metrics emerging in the IoT architecture, such as network QoS, sensing quality, and computation quality, and metaphysical metrics demanded by users, such as accuracy, context and timeliness.
Pal et al. [7] classified the QoE IFs for IoT environments into three distinct categories:
  • Technical, which represent the various QoS factors, which are popular in the multimedia context and also relevant with the IoT examined scenario.
  • User, which represent the subjective characteristics of the users of the IoT applications.
  • Context, which are related to the data and information quality along with specific application requirements that can vary depending upon the usage scenario.
Figure 4 presents the taxonomy for IoT environments by Pal et al. [7].
Nashaat et al. [23] consider 3 dimensions: the Environment runtime context, the Application, and the User expectations. These factors, in addition to QoS feedback, influence the total QoE of the user by a valuable weight, as Figure 5 depicts.
Besides the QoE taxonomies for IoT applications, researchers have proposed various QoE taxonomies for specific IoT verticals. For example, Damaj et al. [37] in their taxonomy for the context of Connected and Autonomous Electric Vehicles (CAEVs) have identified several performance indicators that were grouped into categories. These categories were then mapped to 4 QoE IFs. Table 4 presents the categories and the corresponding QoE IF.
Finally, Wette et al. [38] listed the most common IoT QoE IFs that are used in a generic Machine QoE (M-QoE) framework developed by Ericsson to accurate predict the QoE of the IoT subscribers. The selected IFs constitute different QoS, QoD and QoC metrics.
An overview of the different quality metrics are presented in Table 5.

5. Quality Models for IoTs

Traditionally, qualitative methods that focus on voice perceptibility for applications usability are used for the QoE evaluation [39]. The QoE for multimedia services is evaluated by subjective, objective and hybrid assessment (a combination of both the subjective and objective approaches) [40].
In this context, a few studies that focus on modeling the relationship between human experience and quality perception in relation to the smart-wearable segment may be found in the literature [6,41,42]. QoE is considered a very important aspect of multiple sensorial media (mulsemedia) [43].
Shin et al. [41] examined the relation of users’ experience and the quality perception in IoT. To achieve this goal the authors utilized a combination of qualitative and quantitative methods. Figure 6 shows the proposed QoE model in which, besides the user’s behavior, coolness, satisfaction and affordance are considered as QoE factors in the IoT context.
Pal et al. [44] proposed a QoE model that maps QoD and QoI to QoE. More specifically, the authors, in order to create the model, collect data from 5 wearable devices. Half of the data set is used to build the model, while the other half is used to test accuracy. The step-counts and heart-rate measurement readings by the wearables are used as QoD parameters, whereas the perceived ease of use, perceived usefulness, and richness in information are used as QoI parameters. The accuracy of their model is evaluated by comparing the QoE obtained from the mathematical model and a subjective test with 40 participants. The authors adopted the Mean Opinion Score (MOS) to quantify the user experience.
Saleme et al. [6] studied the impact that human factors such as gender, age, prior computing experience, airflow intensity and smell sensitivity have on 360 mulsemedia QoE. A total of 48 participants (27 male, 21 female) participated in this study with ages between 16 and 65 years old. Results showed that all these factors influence the users’ QoE. Guidelines for evaluating wearables’ quality of experience in a mulsemedia context can be found in [43].
In addition to the QoE evaluation for wearables, several attempts have been made in order to create QoE models for IoT. One of the first attempts was made by Wu et al. [45] that calculated the overall QoE by combining two parameters: profit (expressed in terms of QoD, QoI, QoE) and cost (expressed in terms of resource efficiency, i.e., device utilization efficiency, computational efficiency, energy efficiency, storage efficiency). The same approach was also followed in several other studies illustrated in Table 6.
Another way of quantifying QoE is the layer-based approach [23], in which each layer focuses on a specific QoE IF (domain), so that the overall quality can be computed as a combination of all IFs (domains). Several layered-QoE models may be found in the related literature.
For example, Floris and Atzori [20] proposed a layered QoE model that aims to evaluate the contributions of each IF to estimate the overall QoE in Multimedia IoT (MIoT) applications [23]. More specifically, the proposed model consists of five layers: physical devices, network, combination, application, and context. The authors, in order to demonstrate the generalization of the their framework, have applied it in two use cases: (a) vehicles remote monitoring and (b) smart surveillance.
A similar approach is also presented in [36]. More specifically, in this framework, the physical metrics are organized into four layers (device, network, computing, and user interface) while the metaphysical metrics are organized in two layers (information and comfort). However, no evaluation of the proposed framework is provided in this work.
Suryanegara et al. [46] proposed a different approach to measure the QoE of IoT services. More specifically, their proposed framework is based on the following steps: (1) Setting up the focus of the IoT services to formulate the QoE parameters, (2) Judging the institutional users who are the users of IoT services, (3) Conducting a Mean Opinion Score (MOS) survey of IoT service users, (4) Calculating the differential MOS as the Absolute Category Rating with Hidden References (ACR-HR) quantitative scale, and (5) Providing the strategic implications to those responsible for the implementation. The authors, in order to validate the proposed framework, conducted a subjective test in Jakarta where 6 institutional users expressed their experience of utilizing IoT technology in their relevant services, i.e., managing public transportation, garbage trucks, ambulances, the fire and rescue brigade, street lighting, and water level measurement.
Finally, [17] proposed a framework to measure the quality of autonomic IoT by mapping five IoT quality metrics to the IoT application life cycle stages: (1) Data Sensing, (2) Sensed Data Transmission, (3) Data Analytics, (4) Analyzed Information Transmission, and (5) Actuation. However, approaches on how to model and measure these IoT IFs are still an open issue.

6. Discussion

Defining quality in an IoT environment it is not an easy task. Although several terms are proposed in the literature, with the heterogeneity of the IoT components, it is difficult to have a generic definition for quality in IoT. A specific domain definition seems to be a more appropriate solution as in [17,29]; however, a classification based on the different characteristics of the IoT is required.
Additionally, the diversity of IoT applications makes the identification of the appropriate IFs a very challenging task. In Section 3, we have collected all the Quality IFs that can be found in the literature, as Figure 7 depicts. However, the answer to the question “which IFs should be considered for this IoT application” cannot be easily provided.
Machine Learning (ML) techniques can be beneficial to address this challenge, since they can be applied to predict the type of IoT application and, as a sequence, the appropriate IFs. Saovapakhiran et al. [47] in their proposed QoE-driven IoT architecture propose the use of ML techniques to tailor QoE at the User level from User Engagement metrics. However, they do not provide a quantitative solution.
In addition, the fact that existing IoT architectures are multi-tier systems increases the complexity of the measurements in IoT. Since each tier has different aspects for the quality IFs, it is not often clear where the IFs should be collected. For example, in the M-QoE framework [38], the IFs are measured for: (1) the IoT device, (2) the Radio Access Network (RAN), (3) the edge network, (4) the core network/services network and (5) the vertical slices and service layers. However, acquiring data from different tiers can result in deterioration of communication delays [47]. To deal with this issue, Saovapakhiran et al. [47] suggested the creation of different QoE domains and the local estimation of QoE in each domain. However, no implementation details are given. In addition, the fact there is a standardized architecture the QoE domains may differ according to the proposed architecture.
Furthermore, security and privacy are crucial challenges to be addressed in IoT architectures. For example, in wearable environments, as more data are collected for the QoE evaluation, the more users’ personalized data are revealed. In addition, the multi-tier IoT architectures make security provision difficult. Especially, for vehicular environments in which the topology of the computing network frequent changes due to mobility, security provision is harder to be achieved compared to other networks.
Additionally, quality assessment in IoT requires further research study. Subjective tests are considered the core part of QoE evaluation for multimedia environments. However, existing subjective approaches that are used to measure QoS, may not be suitable for IoT environments, since (1) it is not feasible to carry out subjective tests for every existing or new IoT application due to their big diversity; (2) they require user feedback after every specific interval, resulting in high network delay and relatively low application response time [23]. In particular, for real time monitoring IoT applications, this can lead to malfunction or in other cases can be dangerous even for humans’ safety; (3) they cannot easy determine the cause-roots of the performance, e.g., as stated in [48], subjective results for autonomous vehicles cannot be very helpful for policymakers to define the cause of a car accident; and (4) subjective assessment requires human participation and is usually performed in a (rather isolated) lab environment. Even if we build objective models from subjective tests, their validity will be limited only to the application scenarios for which they are tested [44]. Thus, further study of QoE assessment in IoT is required.
Finally, the conducted research showed that although mulsemedia content provides a new content experience that goes beyond traditional media, QoE evaluation for such types of content is an under-researched area. More QoE IFs should be determined in order to reflect the human-to-machine interaction and, thus, create accurate QoE models. However, the complexity of this task is further increased due to the fact that the majority of the existing olfactory information based systems and methods is only available in specialized laboratories [49]. In addition, there is guidelines on how to create a multisensory content [50], as well as there are not many mulsemedia datasets available.
Table 7 overviews the challenges concerning quality in IoT.

7. Conclusions

Quality is an important factor in an IoT environment. Quality provisioning in such environments is not only limited to life-threatening situations, but also needs to consider the risk of causing significant business losses and environmental damage [48]. QoE is the most popular metric that has been used to evaluate quality. However, due to fact that initially QoE was introduced to assess the end-user satisfaction, the concepts of traditional QoE should be extended in order to include contextual factors that are important in the IoT domain. In addition, more quality metrics are required in order to evaluate quality in IoT. To this extend, this paper has surveyed the actual necessity of evaluating the quality in IoT. We identified the quality metrics that impact the quality in an IoT environment. However, even the collection of the quality metrics measurements is not an easy process. For one point, data should be collected from multiple IoT nodes locating at different tiers depending on the IoT architecture that is used, while the storage and transfer of these obtained large-scale data is a very challenging task. Even existing assessing methods should be re-examined in the context of IoT. Especially for mulsemedia applications, traditional QoE assessment methods are not adequate. Thus, research is needed in order to deal with these challenges.

Author Contributions

Contributions: Writing—original draft: A.S.; resources: A.S. and P.C. Writing—review & editing: P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. ITU-T. Quality of telecommunication services: Concepts, models, objectives and dependability planning–Terms and definitions related to the quality of telecommunication services. In SERIES E: Overall Network Operation, Telephone Service, Service Operation and Human Factors; International Telecommunication Union (ITU): Geneva, Switzerland, 2008. [Google Scholar]
  2. Varela, M.; Skorin-Kapov, L.; Ebrahimi, T. Quality of service versus quality of experience. In Quality of Experience; Springer: Cham, Switzerland, 2014; pp. 85–96. ISBN 978-3-319-02680-0. [Google Scholar]
  3. Chen, C.W.; Chatzimisios, P.; Dagiuklas, T.; Atzori, L. Multimedia Quality of Experience (QoE): Current Status and Future Requirements; John Wiley & Sons: West Sussex, UK, 2015; ISBN-13: 978-1118483916; ISBN-10: 9781118483916. [Google Scholar]
  4. Laghari, K.U.R.; Connelly, K. Toward total quality of experience: AQoE model in a communication ecosystem. IEEE Commun. Mag. 2012, 50, 58–65. [Google Scholar] [CrossRef]
  5. Baraković, S.; Skorin-Kapov, L. Survey and challenges of QoE management issues in wireless networks. J. Comput. Netw. Commun. 2013, 2013, 165146. [Google Scholar] [CrossRef] [Green Version]
  6. Saleme, E.B.; Covaci, A.; Assres, G.; Comsa, I.S.; Trestian, R.; Santos, C.A.; Ghinea, G. The influence of human factors on 360 mulsemedia QoE. Int. J. Hum.-Comput. Stud. 2021, 146, 102550. [Google Scholar] [CrossRef]
  7. Pal, D.; Vanijja, V.; Varadarajan, V. Quality Provisioning in the Internet of Things Era: Current State and Future Directions. In Proceedings of the 10th International Conference on Advances in Information Technology, Bangkok, Thailand, 10–13 December 2018; pp. 1–7. [Google Scholar]
  8. Floris, A.; Atzori, L. Quality of Experience in the Multimedia Internet of Things: Definition and practical use-cases. In Proceedings of the 2015 IEEE International Conference on Communication Workshop (ICCW), London, UK, 8–12 June 2015; pp. 1747–1752. [Google Scholar]
  9. Rahman, T.; Pal, D.; Funilkul, S.; Photikit, N. Apples vs. Oranges: The QoE Scenario in Consumer IoT Services. In Proceedings of the 2022 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 7–9 January 2022; pp. 1–5. [Google Scholar]
  10. Nauman, A.; Qadri, Y.A.; Amjad, M.; Zikria, Y.B.; Afzal, M.K.; Kim, S.W. Multimedia Internet of Things: A comprehensive survey. IEEE Access 2020, 8, 8202–8250. [Google Scholar] [CrossRef]
  11. Kassab, W.; Darabkh, K.A. A–Z survey of Internet of Things: Architectures, protocols, applications, recent advances, future directions and recommendations. J. Netw. Comput. Appl. 2020, 163, 102663. [Google Scholar] [CrossRef]
  12. Taha, A.E.M. Quality of experience in 6G networks: Outlook and challenges. J. Sens. Actuator Netw. 2021, 10, 11. [Google Scholar] [CrossRef]
  13. Skorin-Kapov, L.; Varela, M.; Hoßfeld, T.; Chen, K.T. A survey of emerging concepts and challenges for QoE management of multimedia services. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2018, 14, 1–29. [Google Scholar] [CrossRef]
  14. Barakabitze, A.A.; Barman, N.; Ahmad, A.; Zadtootaghaj, S.; Sun, L.; Martini, M.G.; Atzori, L. QoE management of multimedia streaming services in future networks: A tutorial and survey. IEEE Commun. Surv. Tutor. 2019, 22, 526–565. [Google Scholar] [CrossRef] [Green Version]
  15. Fizza, K.; Banerjee, A.; Mitra, K.; Jayaraman, P.P.; Ranjan, R.; Patel, P.; Georgakopoulos, D. QoE in IoT: A vision, survey and future directions. Discov. Internet Things 2021, 1, 1–14. [Google Scholar] [CrossRef]
  16. Bures, M.; Bellekens, X.; Frajtak, K.; Ahmed, B.S. A comprehensive view on quality characteristics of the IoT solutions. In EAI International Conference on IoT in Urban Space; Springer: Cham, Switzerland, 2018; pp. 59–69. ISBN-13 978-3030289249. [Google Scholar] [CrossRef]
  17. Fizza, K.; Banerjee, A.; Jayaraman, P.P.; Auluck, N.; Ranjan, R.; Mitra, K.; Georgakopoulos, D. A Survey on Evaluating the Quality of Autonomic Internet of Things Applications. IEEE Commun. Surv. Tutor. 2022. [Google Scholar] [CrossRef]
  18. Cichy, C.; Rass, S. An Overview of Data Quality Frameworks. IEEE Access 2019, 7, 24634–24648. [Google Scholar] [CrossRef]
  19. Martinez-Caro, J.M.; Cano, M.D. A novel holistic approach for performance evaluation in Internet of Things. Int. J. Commun. Syst. 2021, 34, e4454. [Google Scholar] [CrossRef]
  20. Floris, A.; Atzori, L. Managing the quality of experience in the multimedia internet of things: A layered-based approach. Sensors 2016, 16, 2057. [Google Scholar] [CrossRef] [Green Version]
  21. Yousefpour, A.; Ishigaki, G.; Gour, R.; Jue, J.P. On reducing IoT service delay via fog offloading. IEEE Internet Things J. 2018, 5, 998–1010. [Google Scholar] [CrossRef] [Green Version]
  22. Khosroabadi, F.; Fotouhi-Ghazvini, F.; Fotouhi, H. Scatter: Service placement in real-time fog-assisted iot networks. J. Sens. Actuator Netw. 2021, 10, 26. [Google Scholar] [CrossRef]
  23. Nashaat, H.; Ahmed, E.; Rizk, R. IoT application placement algorithm based on multi-dimensional QoE prioritization model in fog computing environment. IEEE Access 2020, 8, 111253–111264. [Google Scholar] [CrossRef]
  24. Nayeri, Z.M.; Ghafarian, T.; Javadi, B. Application placement in Fog computing with AI approach: Taxonomy and a state of the art survey. J. Netw. Comput. Appl. 2021, 185, 103078. [Google Scholar] [CrossRef]
  25. Mouradian, C.; Naboulsi, D.; Yangui, S.; Glitho, R.H.; Morrow, M.J.; Polakos, P.A. A comprehensive survey on fog computing: State-of-the-art and research challenges. IEEE Commun. Surv. Tutor. 2017, 20, 416–464. [Google Scholar] [CrossRef] [Green Version]
  26. ITU-T. 10: Vocabulary for performance and quality of service, Amendment 2: New definitions for inclusion in Recommendation ITU-T P. 10/G. 100. Amendment 1 (06/19). 2008. Available online: https://www.itu.int/rec/dologin_pub.asp?lang=s&id=T-REC-P.10-200807-S!Amd2!PDF-E&type=items (accessed on 22 November 2022).
  27. Raake, A.; Egger, S. Quality and quality of experience. In Quality of Experience; Springer: Berlin/Heidelberg, Germany, 2014; pp. 11–33. [Google Scholar] [CrossRef]
  28. Brunnström, K.; Beker, S.; De Moor, K.; Dooms, A.; Egger, S.; Garcia, M.-N.; Hoßfeld, H.; Jumisko-Pyykkö, S.; Keimel, C.; Larabi, C.; et al. Qualinet white paper on definitions of quality of experience. In European Network on Quality 397 of Experience in Multimedia Systems and Services (COST Action IC 1003), hal-00977812, version 1; HAL: Lyon, France, 2012; Volume 3. [Google Scholar]
  29. Minovski, D.; Åhlund, C.; Mitra, K. Modeling quality of IoT experience in autonomous vehicles. IEEE Internet Things J. 2020, 7, 3833–3849. [Google Scholar] [CrossRef]
  30. Karaadi, A.; Sun, L.; Mkwawa, I.H. Multimedia communications in internet of things QoT or QoE? In Proceedings of the 2017 IEEE International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), Exeter, UK, 21–23 June 2017; pp. 23–29. [Google Scholar]
  31. Karaadi, A.; Mkwawa, I.H.; Sun, L. How to manage multimedia traffic: Based on QoE or QoT? ZTE Commun. 2020, 16, 23–29. [Google Scholar]
  32. Wang, Y.; Li, X.; Ji, H.; Zhang, H. QoX-Driven Hierarchical Networking Scheme for Multi-UAV Assisted IoT Networks. In Proceedings of the 2022 IEEE Wireless Communications and Networking Conference (WCNC), Austin, TX, USA, 10–13 April 2022; pp. 120–125. [Google Scholar]
  33. Stankiewicz, R.; Jajszczyk, A. A survey of QoE assurance in converged networks. Comput. Netw. 2011, 55, 1459–1473. [Google Scholar] [CrossRef]
  34. Iggena, T.; Muslim, A.B.; Fischer, M.; Toenjes, R. Quality of Information aware LoRaWAN for mobile Nodes. In Proceedings of the Mobile Communication-Technologies and Applications; 25th ITG-Symposium, Osnabrueck, Germany, 3–4 November 2021; ISBN 978-3-8007-5674-2. [Google Scholar]
  35. Kuemper, D.; Iggena, T.; Toenjes, R.; Pulvermueller, E. Valid.IoT: A framework for sensor data quality analysis and interpolation. In Proceedings of the 9th ACM Multimedia Systems Conference, Amsterdam, Netherlands, 12–15 June 2018; pp. 294–303. [Google Scholar]
  36. Ikeda, Y.; Kouno, S.; Shiozu, A.; Noritake, K. A framework of scalable QoE modeling for application explosion in the Internet of Things. In Proceedings of the 2016 IEEE 3rd World Forum on Internet of Things (WF-IoT), Reston, VA, USA, 12–14 December 2016; pp. 425–429. [Google Scholar]
  37. Damaj, I.W.; Serhal, D.K.; Hamandi, L.A.; Zantout, R.N.; Mouftah, H.T. Connected and autonomous electric vehicles: Quality of experience survey and taxonomy. Veh. Commun. 2021, 28, 100312. [Google Scholar] [CrossRef]
  38. Tchouati, C.W.; Rochefort, S.; Sarmonikas, G. Monitoring IoT application performance with Machine QoE. Ericsson Technol. Rev. 2020, 8, 2–11. [Google Scholar] [CrossRef]
  39. Pierucci, L. The quality of experience perspective toward 5G technology. IEEE Wirel. Commun. 2015, 22, 10–16. [Google Scholar] [CrossRef]
  40. Yang, M.; Wang, S.; Calheiros, R.N.; Yang, F. Survey on QoE assessment approach for network service. IEEE Access 2018, 6, 48374–48390. [Google Scholar] [CrossRef]
  41. Shin, D.H. Conceptualizing and measuring quality of experience of the internet of things: Exploring how quality is perceived by users. Inf. Manag. 2017, 54, 998–1011. [Google Scholar] [CrossRef]
  42. Pal, D.; Vanijja, V.; Arpnikanondt, C.; Zhang, X.; Papasratorn, B. A quantitative approach for evaluating the quality of experience of smart-wearables from the quality of data and quality of information: An end user perspective. IEEE Access 2019, 7, 64266–64278. [Google Scholar] [CrossRef]
  43. Hussain, N.; Ghinea, G. Guidelines for evaluating wearables’ quality of experience in a mulsemedia context. Multimed. Tools Appl. 2022, 81, 43283–43314. [Google Scholar] [CrossRef]
  44. Pal, D.; Triyason, T.; Varadarajan, V.; Zhang, X. Quality of experience evaluation of smart-wearables: A mathematical modelling approach. In Proceedings of the 2019 IEEE 35th International Conference on Data Engineering Workshops (ICDEW), Macao, China, 8–12 April 2019; pp. 74–80. [Google Scholar]
  45. Wu, Q.; Ding, G.; Xu, Y.; Feng, S.; Du, Z.; Wang, J.; Long, K. Cognitive internet of things: A new paradigm beyond connection. IEEE Internet Things J. 2014, 1, 129–143. [Google Scholar] [CrossRef]
  46. Suryanegara, M.; Prasetyo, D.A.; Andriyanto, F.; Hayati, N. A 5-step framework for measuring the quality of experience (qoe) of internet of things (iot) services. IEEE Access 2019, 7, 175779–175792. [Google Scholar] [CrossRef]
  47. Saovapakhiran, B.; Naruephiphat, W.; Charnsripinyo, C.; Baydere, S.; Ozdemir, S. QoE-Driven IoT Architecture: A Comprehensive Review on System and Resource Management. IEEE Access 2022. [Google Scholar] [CrossRef]
  48. Minovski, D.; Åhlund, C.; Mitra, K.; Zhohov, R. Defining quality of experience for the internet of things. IT Prof. 2020, 22, 62–70. [Google Scholar] [CrossRef]
  49. Akhtar, Z.; Siddique, K.; Rattani, A.; Lutfi, S.L.; Falk, T.H. Why is multimedia quality of experience assessment a challenging problem? IEEE Access 2019, 7, 117897–117915. [Google Scholar] [CrossRef]
  50. Covaci, A.; Saleme, E.B.; Mesfin, G.; Hussain, N.; Kani-Zabihi, E.; Ghinea, G. How do we experience crossmodal correspondent mulsemedia content? IEEE Trans. Multimed. 2019, 22, 1249–1258. [Google Scholar] [CrossRef]
Figure 1. An IoT applications’ classification.
Figure 1. An IoT applications’ classification.
Iot 03 00026 g001
Figure 2. The fog computing architecture.
Figure 2. The fog computing architecture.
Iot 03 00026 g002
Figure 3. Factors influencing QoE [33].
Figure 3. Factors influencing QoE [33].
Iot 03 00026 g003
Figure 4. QoE taxonomy for IoT [7].
Figure 4. QoE taxonomy for IoT [7].
Iot 03 00026 g004
Figure 5. QoE IFs in the IoT context [23].
Figure 5. QoE IFs in the IoT context [23].
Iot 03 00026 g005
Figure 6. QoE IFs in the IoT context [7].
Figure 6. QoE IFs in the IoT context [7].
Iot 03 00026 g006
Figure 7. Quality IFs in the IoT context.
Figure 7. Quality IFs in the IoT context.
Iot 03 00026 g007
Table 1. Quality and IoT.
Table 1. Quality and IoT.
PaperRecipientTermShortcoming
UserMachine
[9] xQoT  Too generic definition.
It is not clear how it can be measured
[17]xxQoE AIoT Autonomic IoT systems
[26]x QoEIt does not reflect the machine’s focused quality
[27]x QoEIt does not reflect the machine’s focused quality
[28]x QoEIt does not reflect the machine’s focused quality
[29]xxQoIoTIt cannot be applied to Autonomic IoT systems
[30]xxQoE IoT It cannot be applied to Autonomic IoT systems
Table 2. QoD metrics.
Table 2. QoD metrics.
MetricDefinition
CompletenessThe extent to which data are of sufficient breadth,
depth and scope for the task at hand [18]
PrecisionThe extent to which the collected data are precise
TruthfulnessThe extent to which the collected data are from reliable resource [19]
AccuracyThe extent to which data are correct and accepted
UsefulnessThe extent to which the sensed data are for the application [15]
ConsistencyThe extent to which data are presented in
the same format and compatible with previous data [18]
TimelinessThe extent to which data are valid for decision making [15]
Table 3. QoI metrics.
Table 3. QoI metrics.
MetricDefinition
RecallProportion of relevant information retrieved from a query [9]
DetailCompleteness of the information provided to the decision-maker [9]
ValidityProvided information is true or not [19]
AccuracyAccuracy degree of information to the decision maker [19]
TimelinessTimely information for an IoT service (opposite to latency) [19]
PrecisionHow close the measured values are to each other [19]
Table 4. QoE IFs for CAEVs [6].
Table 4. QoE IFs for CAEVs [6].
QoE IFCategories
ContextTravel Efficiency, Operability
CostAffordability
QoSEnergy, Security, Networking and Connectivity, Survivability, Subsystem Performance
HumanSafety, Personal Usability
Table 5. Quality IFs for IoT.
Table 5. Quality IFs for IoT.
Quality IFDescriptionRefs.
Quality of Experience (QoE)/
Human feedback
Evaluates the overall acceptability of an application or service or system as perceived subjectively by users[7,15,30,38]
Quality of Context (QoC)Evaluates the context of the environment or the application[7,17,23]
Quality of Cost (QoCo)Evaluates the cost in terms of of computation, storage, or energy of an IoT application[9]
Quality of Information (QoI)Evaluates the quality of information[9,15,17,19,34]
Quality of Data (QoD)Evaluates the quality of data[15,17,18,19]
Quality of Service (QoS)Evaluates the network’s capability to provide satisfied service levels[15,17,23,33,38]
Quality of Device (QoDe)Evaluates the quality of the physical IoT devices[17]
Quality of Actuation (QoA)Evaluates the correctness of the decision making/ actuation performed by an IoT application[17]
Quality of Security and Privacy
(QoSe & P)
Evaluates the security and privacy of an IoT application[17]
Table 6. Generic QoE calculations comparison.
Table 6. Generic QoE calculations comparison.
PaperQoEQoDQoIQoCTesting
[9]Evaluated from factors of 3 level- architecture (access, communication, computation, application)Accuracy, truthfulness, completeness, up-to-dateness, precisionTimeliness, validity, recall, accuracy, detailComputational efficiency, energy efficiency, storage efficiencyNo
[19]Delay, jitter, packet delivery rate, throughput, and gateway availability.Completeness, precision, truthfulnessQuality, precision, recall, accuracy, detail, timeliness, validityEnergy consumption, interface useSimulation
[45]Evaluated from factors of 4 level- architecture (access, communication, computation)Accuracy, truthfulness, completeness, up-to-dateness, quantity, precisionRecall, accuracy, detail, timeliness, validityDevice utilization efficiency, computational efficiency, energy efficiency, storage efficiencyNo
Table 7. Quality challenges in IoT.
Table 7. Quality challenges in IoT.
ChallengeExisting SolutionDrawback
Define end-to-end quality in IoTUse existing QoE definition Define new terms according to the context of the applicationIt cannot be applied to autonomic IoT Impractical due to the diversity of applications
Identify the appropriate IoT IFsBased on the application different IFs are consideredImpractical due to the diversity of applications
Collect IFs MeasurementsAt different layers from different nodes of the IoT architectureIncrease communication delays There is no unified and inter-operable standard
Security and privacyBased on the application domain different approaches are proposedImpractical due to the diversity of applications and the increasing appearance of new threats
Assess qualitySubjective methods Objective methodsSubjective methods suffer from user bias Objective methods expensive
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sgora, A.; Chatzimisios, P. Defining and Assessing Quality in IoT Environments: A Survey. IoT 2022, 3, 493-506. https://doi.org/10.3390/iot3040026

AMA Style

Sgora A, Chatzimisios P. Defining and Assessing Quality in IoT Environments: A Survey. IoT. 2022; 3(4):493-506. https://doi.org/10.3390/iot3040026

Chicago/Turabian Style

Sgora, Aggeliki, and Periklis Chatzimisios. 2022. "Defining and Assessing Quality in IoT Environments: A Survey" IoT 3, no. 4: 493-506. https://doi.org/10.3390/iot3040026

APA Style

Sgora, A., & Chatzimisios, P. (2022). Defining and Assessing Quality in IoT Environments: A Survey. IoT, 3(4), 493-506. https://doi.org/10.3390/iot3040026

Article Metrics

Back to TopTop