Next Article in Journal
Multivariate Analysis of Difference Raman Spectra of the Irradiated Nucleus and Cytoplasm Region of SH-SY5Y Human Neuroblastoma Cells
Previous Article in Journal
Enabling Heart Self-Monitoring for All and for AAL—Portable Device within a Complete Telemedicine System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Number of Message Transmissions for Probabilistic Guarantee of Latency in the IoT

by
Pascale Minet
*,† and
Yasuyuki Tanaka
*,†
Inria Research Center of Paris, 75012 Paris, France
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2019, 19(18), 3970; https://doi.org/10.3390/s19183970
Submission received: 1 August 2019 / Revised: 6 September 2019 / Accepted: 11 September 2019 / Published: 14 September 2019
(This article belongs to the Section Internet of Things)

Abstract

:
The Internet of Things (IoT) is now experiencing its first phase of industrialization. Industrial companies are completing proofs of concept and many of them plan to invest in automation, flexibility and quality of production in their plants. Their use of a wireless network is conditioned upon its ability to meet three Key Performance Indicators (KPIs), namely a maximum acceptable end-to-end latency L, a targeted end-to-end reliability R and a minimum network lifetime T. The IoT network has to guarantee that at least R % of messages generated by sensor nodes are delivered to the sink with a latency ≤L, whereas the network lifetime is at least equal to T. In this paper, we show how to provide the targeted end-to-end reliability R by means of retransmissions to cope with the unreliability of wireless links. We present two methods to compute the maximum number of transmissions per message required to achieve R. M F a i r is very easy to compute, whereas M O p t minimizes the total number of transmissions necessary for a message to reach the sink. M F a i r and M O p t are then integrated into a TSCH network with a load-based scheduler to evaluate the three KPIs on a generic data-gathering application. We first consider a toy example with eight nodes where the maximum number of transmissions M a x T r a n s is tuned per link and per flow. Finally, a network of 50 nodes, representative of real network deployments, is evaluated assuming M a x T r a n s is fixed. For both TSCH networks, we show that M O p t provides a better reliability and a longer lifetime than M F a i r , which provides a shorter average end-to-end latency. M O p t provides more predictable end-to-end performances than Kausa, a KPI-aware, state-of-the-art scheduler.

1. Introduction

The Internet of Things (IoT) is transforming our daily life at home [1], at work, in our cities [2], in transportation [3], in sport training and healthcare, in process control and automation, in smart farming and beyond.

1.1. Context

At home or in the office, the IoT allows us to save time and energy by controlling lights and appliances and knowing our resource consumption habits. In business and industry, it increases productivity and efficiency by streamlining processes. In transportation, it helps people to enjoy services of better quality. IoT devices are electronic devices able to communicate with a network and perform a task. We can draw a distinction between consumer devices, smart home devices, enterprise and industrial IoT devices. In this paper, we do not focus on IoT devices themselves, but rather on applications using these devices. More precisely, we will consider IoT applications that are responsible for data gathering and are the most demanding in terms of quality of service provided by the networks supporting them [4,5]. We will consider three types of requirements:
  • end-to-end reliability, denoted R: the data generated by any sensor node must be delivered to the sink with a probability greater than or equal to R. A value of 0.99 is frequent. In industry, a value of 0.99999 is targeted.
  • end-to-end latency, denoted L: since, on the one hand, the role of the network is to collect data generated by wireless sensor nodes that are deployed in the area covered by this network and, on the other hand, only up-to-date information can be used to take accurate decisions, the time elapsed between data generation and data delivery to the sink is defined as the end-to-end latency. A value less than one second is usual.
  • network lifetime, denoted T: since most IoT devices are battery operated, the goal of the network is to maximize its lifetime defined as the time when the first device has exhausted its battery. The importance of this requirement increases with the difficulty or the cost of replacing a battery in the environment concerned. A lifetime of several years (e.g., 3 years) is usual.
Since most wireless sensor networks deployed up to now are based on the IEEE 802.15.4 technology [6] and, given that the TSCH (Time Slotted Channel Hopping) technology [7] has been designed to enhance the IEEE 802.15.4 technology to better meet the three types of requirements previously defined, we focus our study on TSCH networks. Is a TSCH network able to meet the usual requirements of IoT applications? The goal of this paper is to answer this question, which can be refined into: is a TSCH network capable of ensuring that at least R percent of messages will be delivered to the sink with a delay less than or equal to L, and a network lifetime greater than or equal to T, while taking into account the unreliability of wireless links.

1.2. Contributions

Before going further, we define the concept of flow. A f l o w is defined as a set of messages having the same attributes such as the source address, the destination address and the QoS (Quality of Service) parameters (e.g., end-to-end reliability, end-to-end latency).
In contrast to many network performance evaluations, we take into account the unreliability of wireless links. The first contribution of this paper is to show how to reach the targeted end-to-end reliability R with two methods M F a i r and M O p t able to estimate the maximum number of transmissions per message and per flow over each link visited, taking into account both the heterogeneity of link quality and the dynamicity of link quality observed on real traces. Whereas  M F a i r is very simple, M O p t minimizes the total number of transmissions per message, for each flow considered. The second contribution consists of providing a fully-integrated approach for a TSCH network by integrating these methods and a Load-based scheduler in the 6TiSCH stack. As a third contribution, we evaluate the performances (i.e., the end-to-end reliability and the end-to-end latency) obtained by this approach on two TSCH networks, one of eight nodes and the other of 50 nodes. Finally, we show that, on a generic data-gathering application, this approach provides more predictable end-to-end performances than K a u s a [8], a KPI-aware state-of-the-art scheduler.
This paper is organized as follows. In Section 2, we present the advantages and drawbacks of well-known link quality estimators and show how they are used to enhance network performances. Since the approach proposed in this paper shares the same goals as K a u s a [8], K a u s a is briefly described. A more detailed description is given in Section 7. Section 3 describes the two methods we propose to estimate the maximum number of transmissions per message and per flow over each link visited. Section 4 shows how to integrate these methods in a TSCH network with a centralized scheduling function. Section 5 gives performance results with regard to end-to-end reliability and end-to-end latency for a targeted end-to-end reliability ranging from 0.9 to 0.99999 for a toy example. In Section 6, a TSCH network of 50 nodes is considered with a generic data-gathering application. End-to-end latency and reliability are evaluated. The solution proposed is compared to K a u s a in Section 7. Finally, we conclude in Section 8.

2. Related Work

Different ways to estimate the quality of a wireless link exist. They all rely on link measurements obtained by an active, passive or hybrid link monitoring. A usual classification distinguishes between hardware and software-based link quality estimators [9,10]. Hardware estimators have the advantage of being directly provided by hardware without the need for any additional processing overhead. However, their accuracy is not very good for two reasons. First, they are measured only on successfully received packets, and they do not take into account the number of packet losses. Second, they are not evaluated on the whole packet received but only on eight symbols of this packet. The main hardware-based quality estimators are RSSI (Received Signal Strength Indicator), LQI (Link Quality Indicator), and SNR (Signal-to-Noise Ratio). In addition, LQI depends on the manufacturer of the radio transceiver used. For the software-based quality estimators, we distinguish those based on PRR (Packet Reception Rate), from those based on RNP (Required Number of Packet retransmissions) and finally those using a score. For instance, ETX (Expected Transmission Count) [11] is classified as an RNP-based estimator, whereas F-LQE (Fuzzy Link Quality Estimator) [12] is a score-based one assessing link quality in terms of four properties: Smoothed Packet reception Ratio, Stability factor, ASymmetry Level and channel Average Signal-to-Noise Ratio. In [13], the authors introduce B m a x, the maximum count of consecutive transmissions needed for a frame to be transmitted on a link. This metric evaluates the link burstiness, where a burst is defined as a period of continuous packet loss. It is obtained empirically for each link based on previous transmissions and can be seen as an observed worst case. Then, they build a schedule that allocates B m a x cells for every frame that a link is supposed to carry. The problem with such a method is that it may require a long delay (more than 140 hours according to the authors) before a representative B m a x is known for all the links.
Whatever the link quality estimator used, authors usually agree on the following conclusions:
  • The quality of any wireless link can be classified as good, intermediate or bad. The challenge lies in an accurate estimation of the intermediate link quality.
  • An ideal link quality estimator should be [10] energy efficient (i.e., requiring low processing, communication and memory overhead), accurate (i.e., reflecting the real link behavior), reactive (i.e., able to promptly react to persistent link state changes) and stable (i.e., able to tolerate transient link state changes).
  • To better reflect the real behavior of a link, several link properties should be taken into account. That is why link quality estimators tend to combine several simple estimators [14] and use sophisticated techniques (e.g., simple average, filtering, machine learning [15,16], regression, and fuzzy logic [17,18]) to produce a metric from link measurements.
Link quality estimators have been used to improve the quality of service provided to end-users [19]. By using links of better quality, the network throughput is maximized, the delivery times are minimized, routes are more stable [18], etc. These improvements can be increased by link quality prediction using online machine learning techniques as in [20]. In such a case, packets are then able to avoid links before their quality degrades below an acceptable threshold, and routing reactivity is improved. Furthermore, if both the link quality estimator and the routing protocol take energy into account, network lifetime is maximized, as in [21].
Closer to our work, Gaillard et al. propose a greedy algorithm that optimizes the distribution of links used by flows network-wide [22]. These authors extend TASA [23], a well-known centralized scheduling algorithm for FTDMA networks, to take into account retransmissions and fragmented packets. They build a schedule that complies with reliability expectation by adding extra cells that are used in the case of consecutive retransmissions. They take into account link quality and packet fragmentation. The same authors further extend their research in [8] by proposing Kausa, a centralized scheduling algorithm that builds resource paths that guarantee QoS per flow, when multiple applications are using the same network, and each application has its own requirements and traffic flows.
In this paper, we propose to use a simple link quality estimator: PDR (Packet Delivery Rate) that is easily computed by the sender as the ratio of the number of acknowledged frames to the number of sent frames. Even if this estimator is not ideal, as previously defined, it provides an accurate estimation of link behavior because, unlike hardware-based link estimators, it takes into account packet losses. We use the PDR link quality estimator to compute the maximum number of transmissions per message and per flow over any link visited by the flow considered.
We share the same goal as K a u s a [8]: namely, to build a centralized schedule ensuring that flows meet their required end-to-end latency and delivery rate, by means of message retransmissions. Like [8], we adopt a per flow approach, allowing us to differentiate QoS per flow. However, unlike [8], we consider the optimal solution for any flow to be the one that minimizes the total number of transmissions per message of this flow, while achieving the desired end-to-end reliability. Furthermore, traces obtained from real IoT networks like [24,25] allow us to better understand the challenges of wireless networking and make realistic assumptions. That is why, in this paper, the performance evaluation is based on PDR values computed from a realistic model, as explained in Section 6.1.

3. Optimal Retransmission Estimation

To cope with the unreliability of network links, unacknowledged messages are retransmitted up to a maximum number of retransmissions. We have to determine this maximum number that of course depends on the unreliability of the link considered as well as the targeted end-to-end reliability R.
Since each flow may request its own QoS, we adopt a per-flow approach. Without loss of generality, we assume that each message is labeled with its flow tag, which includes the origin node of the flow.

3.1. Assumptions and Basic Properties

Let us consider the flow f originating from any sensor node N h with h [ 1 , n ] . Let h denote the path length (i.e., the number of hops) from N h to the sink N 0 . To simplify the notation, we assume that the path is defined by N h N h 1 N h 2 N 0 .
For any link j [ 1 , h ] of f’s path, we denote by M j the maximum number of transmissions of any message of f transmitted by node N j to its parent in the routing tree, and P j the probability of successful acknowledgment receipt after a single transmission, whereas R j is the probability of successful acknowledgment receipt on link j after a maximum number M j 1 of transmissions. Table 1 gives the notations used in this paper.
For the sake of simplicity, we adopt the following assumptions:
Assumption 1.
Network links fail independently.
Assumption 2.
The transmissions of any message fail independently.
Assumption 3.
Each node has enough Transmission opportunities to prevent message drops due to Transmission queue overflow.
In a TSCH network, this assumption means that the scheduling algorithm has assigned enough transmission cells to each node. As a consequence, the unreliability of a transmission is only due to the unreliability of the wireless link considered. If this assumption were not true, then the computation of the transmission reliability should include not only the message loss due to the unreliability of the wireless link considered but also the message loss due to Transmission queue overflow.
We first recall some basic properties:
Property 1.
The end-to-end reliability on a path is equal to the product of the reliability of each link composing that path.
Property 2.
The probability of successful acknowledgment receipt over any link j increases with the maximum number of message transmissions M j according to the following equation:
R j = 1 ( 1 P j ) M j .
Proof. 
After M j transmissions, the acknowledgment of a message is not received successfully with a probability equal to ( 1 P j ) M j . Hence, the probability of successful acknowledgment receipt after M j transmissions is equal to 1 ( 1 P j ) M j .  □

3.2. A Fair Method

The question is: knowing the end-to-end reliability that must be met on a given path, how to distribute this targeted end-to-end reliability into the targeted reliability of each link on that path? The maximum number of transmissions on each link is then deduced from the targeted reliability on the link and the probability of successful transmission on that link. In this paper, we propose two methods that can be applied to any network meeting Assumptions 1–3.
Since we adopt a per-flow approach, we consider any flow f, originating from a node h hops away from the sink. In other words, the path of f consists of h links, h 1 . Since the end-to-end reliability on the path is equal to the product of the reliability of each of its links, which should be R , a simple solution consists of fairly and uniformly sharing the end-to-end reliability over each link of the path. Hence, each link j has to meet a reliability equal to R j = R 1 / h . The maximum number of transmissions on any link j is then equal to M j = L o g ( 1 R 1 / h ) L o g ( 1 P j ) . This principle is applied by the M F a i r Algorithm given hereafter (see Algorithm 1).
Algorithm 1: M F a i r : compute the number of transmissions on each link j to reach a reliability
R 1 / h over this link per message of flow f visiting h links.
Require:R the targeted end-to-end reliability, P j = the success probability of receiving the
 acknowledgment after a single message transmission over link j with 1 j h
Ensure: M j is the minimum number of transmissions on link j to achieve R 1 / h
for each link j do
  if P j = 1 then
    M j 1
  else
    M j L o g ( 1 R 1 / h ) L o g ( 1 P j )
  end if
end for
Property 3.
The processing complexity of the M F a i r Algorithm is in O ( 1 ) .
Proof. 
The complexity of the Fair Algorithm is equal to the complexity of computing L o g ( 1 R 1 / h ) L o g ( 1 P j ) for each link, times h the number of links, hence the property. □
Figure 1 depicts the maximum number of transmissions over a link whose probability of successful acknowledgment receipt ranges from 0.5 to 0.9 for a 4-hop flow when the targeted end-to-end reliability ranges from 0.9 to 0.9999.

3.3. An Optimal Method

For any sensor node N i , let us denote f the flow originating from N i . The problem consists of minimizing the total number of transmissions needed to deliver any message of f to the sink, under the constraint that the end-to-end latency is greater than or equal to R.
The optimization problem can be defined as follows:
Goal: Find the values of M j for each link j visited by f, j = 1 h that minimize j = 1 h M j under the constraints:
j = 1 h ( 1 ( 1 P j ) M j ) R ,
M j integer 1 .
If several solutions exist for the same total number of transmissions, choose the solution maximizing j = 1 h ( 1 ( 1 P j ) M j ) .
Lemma 1.
Each link j = 1 h should provide a reliability R j at least equal to the requested end-to-end reliability R, R j R . Hence, M j L o g ( 1 R ) L o g ( 1 P j ) .
Proof. 
If there is a link j that provides a reliability R j < R , then the product of the reliability of all other links should be greater than 1 to obtain an end-to-end reliability of R, which is impossible. Therefore, each link has to provide a reliability at least equal to R by means of retransmissions. Since the reliability of a link j increases with the maximum number of transmissions according to Equation (1), to obtain an end-to-end reliability of R, the number of transmissions M j should be greater than or equal to that needed to obtain a link reliability of R, leading to M j L o g ( 1 R ) L o g ( 1 P j ) .  □
Lemma 2.
There is no solution ensuring an end-to-end reliability R , with a total number of transmissions < j = 1 h L o g ( 1 R ) L o g ( 1 P j ) .
Proof. 
It is deduced from Lemma 1. □
Lemma 3.
If there is no solution ensuring an end-to-end reliability greater than or equal to R, with a total number of transmissions M = j = 1 h L o g ( 1 R ) L o g ( 1 P j ) and, if increasing M j to M j + 1 for the link j maximizing the value of P j ( 1 / R j 1 ) does not achieve R, then there is no solution ensuring an end-to-end reliability R in M + 1 transmissions.
Proof. 
Assuming that there is no solution ensuring an end-to-end reliability R , with a total number of transmissions M = j = 1 h L o g ( 1 R ) L o g ( 1 P j ) , we compute, for any link k, the benefit on the end-to-end reliability brought by increasing M k to M k + 1 . The new end-to-end reliability can be written n e w R = ( j = 1 , j k h R j ) ( 1 ( 1 P k ) M k + 1 ) . Since 1 ( 1 P k ) M k + 1 = 1 ( 1 P k ) ( 1 P k ) M k = R k + P k ( 1 R k ) , we get n e w R = ( j = 1 , j k h R j ) ( R k + P k ( 1 R k ) ) . Denoting j = 1 h R j as o l d R , we get n e w R = o l d R ( 1 P k + P k / R k ) . Hence, maximizing n e w R means selecting the link k that maximizes P k ( 1 / R k 1 ) . Hence, if with that link k, the end-to-end reliability n e w R is not greater than or equal to R, then no other link can meet the end-to-end reliability R with a total number of transmissions equal to M + 1 . □
Property 4.
The M O p t algorithm given in Algorithm 2 finds the optimal solution.
Proof. 
The algorithm starts with the smallest possible number of transmissions M, according to Lemma 2. If the requested end-to-end reliability is met, then the solution is found. Otherwise, the algorithm increases the total number of transmissions by one and checks again whether the requested end-to-end reliability is met by increasing the number of transmissions of one on the link providing the highest reliability gain. If yes, the solution is found, otherwise there is no solution for M + 1 transmissions. □
Algorithm 2: M O p t : Compute the number of transmissions M j on link j to achieve an end-to
end reliability R and minimize the total number of transmissions per message of flow f
visiting h links
Require:R the targeted end-to-end reliability 0 < R < 1 , P j =the probability of successful
 acknowledgment receipt after a single transmission over link j with 1 j h
Ensure: M j is the minimum number of transmissions to achieve R
for each link j do
  if P j = 1 then
    M j 1
  else
    M j L o g ( 1 R ) L o g ( 1 P j )
    R j 1 ( 1 P j ) M j
    G a i n j P j ( 1 / R j 1 )
  end if
end for
while j = 1 h R j < R do
  Select the link j maximizing G a i n j
  If several links provide the same G a i n , take the farthest link j from the sink
   M j M j + 1
   R j 1 ( 1 P j ) M j
   G a i n j P j ( 1 / R j 1 )
end while
We now evaluate the complexity of this algorithm and more particularly we upper bound the number of iterations needed to find the maximum number of transmissions over each link such that the total number of transmissions per message of the flow considered is minimized.
Property 5.
The M O p t Algorithm, given in Algorithm 2, finds the optimal solution in a number of iterations less than or equal to j = 1 h L o g ( h ) L o g ( 1 P j ) .
Proof. 
Let us consider any link j visited by the flow considered. Since R 1 , we have R 1 / h < 1 1 R h . Hence, L o g ( 1 R 1 / h ) > L o g ( 1 R ) L o g ( h ) , leading to: L o g ( 1 R 1 / h ) L o g ( 1 P j ) < L o g ( 1 R ) L o g ( h ) L o g ( 1 P j ) . We then obtain L o g ( 1 R 1 / h ) L o g ( 1 P j ) L o g ( 1 R ) L o g ( h ) L o g ( 1 P j ) . Since a b a b , we get: M j , f a i r M j , i n i t L o g ( h ) L o g ( 1 P j ) , where M j , f a i r denotes the maximum number of transmissions over link j, whereas M j , i n i t denotes the first value tried by the optimal algorithm. Since j = 1 h M j , f a i r is the maximum number of transmissions of a message to reach the sink provided by the Fair algorithm and this number is never exceeded by the Optimal algorithm, the maximum number of iterations of the Optimal algorithm is upper bounded by j = 1 h M j , f a i r j = 1 h M j , i n i t L o g ( h ) L o g ( 1 P j ) , hence the property. □

3.4. Expected vs. Maximum Number of Transmissions

For any flow considered and whatever the method adopted to compute the maximum number of transmissions over each link visited by the flow considered, the real number of transmissions used is much smaller than the maximum one, which occurs with a very low probability. In this section, we want to evaluate the energy saving obtained with a variable number of transmissions instead of always considering the worst case that has a very low occurrence probability. With the Assumptions 1–3, we can evaluate E ( M j ) the expected number of transmissions on any link j, knowing that M j , the maximum number of transmissions, is computed either by M F a i r or by M O p t . We then have:
E ( M j ) = k = 1 M j 1 k P j ( 1 P j ) k 1 + M j ( 1 P j ) M j 1 .
Property 6.
The expected number of transmissions on any link j is equal to
E ( M j ) = 1 ( 1 P j ) M j P j ,
where M j denotes the maximum number of transmissions used for link j.
Proof. 
Let us consider S n ( x ) = k = 1 n ( 1 + x ) k and let S n ( x ) be its derivative. We have S n ( x ) = k = 1 n k ( 1 + x ) k 1 . Since S n ( x ) is the sum of a geometric progression, we have:
S n ( x ) = 1 + x ( 1 + x ) n + 1 x = ( 1 + x ) ( ( 1 + x ) n 1 ) x .
Hence, the derivative is:
S n ( x ) = ( 1 + x ) n ( n x 1 ) + 1 x 2 . By replacing x by P j and n by M j 1 , we get
E ( M j ) = k = 1 M j 1 k P j ( 1 P j ) k 1 + M j ( 1 P j ) M j 1 = 1 ( 1 P j ) M j 1 ( 1 ( M j 1 ) P j ) + 1 P j = 1 ( 1 P j ) M j P j , hence the property. □
Figure 2 depicts the maximum and the expected numbers of transmissions per message of a 4-hop flow, for a targeted end-to-end reliability of 0.9999 , over a link whose success probability per transmission ranges from 0.5 to 0.9 . We observed that, in all the cases studied, although the value of E ( M j ) decreases when P j , the success probability per transmission increases, E ( M j ) remains constant and equal to 2. By stopping its retransmissions as soon it receives an acknowledgment, the sender saves its energy. For instance, let us consider that for any link j for which P j = 0.5 is visited by a 4-hop flow, the sender saves M j E ( M j ) transmissions for this link that is 16 2 = 14 transmissions for M F a i r . Since the flow visits four hops and assuming that its four links have the same success probability, each of the four transmitting nodes saves 14 transmissions. Assuming that the schedule is periodic with a period of 101 time slots (the default value in the TSCH network), and this flow generates a message per schedule period, these four nodes save 140 ms each schedule period of 101 × 7.25 = 732.25 ms, for a slot duration of 7.25 ms. This corresponds to an increase of 19 % in sleeping time per schedule period for each node.
It is useful to compare this value with the value of ETX on link j, denoted by E T X j . Let D j be the delivery rate from node j to its parent, whereas D r ( j ) is the delivery rate in the reverse direction. Since P j is defined as the probability for j to receive the acknowledgment of its message, we have to take into account the delivery rates of both directions, which gives:
P j = D j * D r ( j )
E T X j is defined as:
E T X j = 1 D j * D r ( j ) .
As a consequence, E T X j does not depend on the targeted end-to-end reliability R but only on the reliability of both directions of the link considered. This explains the main difference between E T X j and M j .
In addition, E T X j is not equal to E ( j ) . Indeed, E ( j ) assumes a maximum number of transmissions equal to M j , whereas E T X j assumes a number of transmissions that may be infinite.

4. Framework for a TSCH Network

The M O p t and M F a i r methods are now applied to compute the maximum number of transmissions per link and per flow. Flows are generated by a low-power network based on the TSCH technology [7].
We focus on data gathering applications with end-to-end requirements in terms of reliability and latency, as well as requirements with regard to network lifetime. The network supporting these applications is a TSCH network [7].

4.1. TSCH Network

In a TSCH network, the medium access is time-slotted and several transmissions are done on different channels in the same time slot. More precisely, transmissions are scheduled in cells, where a cell is defined by its channel offset and its time slot offset. There are two types of cells: shared cells where any node having a message to transmit is allowed to do so, and dedicated cells, where only the transmitter defined in the schedule is allowed to. The choice of a wireless TSCH network helps to meet Assumptions 1–3 because the mapping between logical channels and physical ones changes at each time slot. Thus, even if a message is retransmitted in the next slot and on the same logical channel as previously, it will be transmitted on a different physical channel.
The schedule of transmissions is periodic and conflicts in dedicated cells are avoided. In addition, nodes know from the schedule in which slots they are allowed to transmit or to receive. They sleep in any other slot in order to save energy.

4.2. Scheduling Function

The scheduling algorithm, which is assumed to be centralized in this paper, works per flow: it allocates the cells needed to transmit a message from the flow origin to the sink. More precisely, it proceeds hop by hop, starting from the flow origin and allocating to each visited node the number of cells needed to receive the message from its child and then the number of cells needed to transmit this message to its parent. Since the scheduler does not know a priori which message transmission will be successful, it has to take into account the worst case where a message is received by the next hop after the maximum number of transmissions for this link and this flow. Hence, for each message, the scheduler allocates to each visited link a number of cells corresponding to the maximum number of transmissions on that link for the flow in question.
However, this does not mean that each message is transmitted a maximum number of times. In fact, as soon as the sender has received the acknowledgment of any message m s g , it stops retransmitting m s g and may use the slot foreseen for a retransmission of m s g for the transmission of another message, if it has one in its T r a n s m i t queue.
Any sensor node first transmits the message in its T r a n s m i t queue that has the highest flow priority, as the primary criterion and the smallest timestamp within a same flow, as the secondary criterion. This assumes that messages are timestamped when they are generated by their origin node.
The Load-based scheduler is selected because of its simplicity combined with its very good performances [26]. This scheduler schedules first the flow originating from the most loaded node. The load of a node is computed as the number of cells needed to transmit its own flows, plus the number of cells needed to receive and transmit the flows originating from its descendants.

4.3. Computation of Key Performance Indicators

In this paper, we consider three Key Performance Indicators (KPIs) that matter for Industry 4.0 and the IoT. We now show how to compute them for a TSCH network and a scheduling function defined in Section 4.1 and Section 4.2, respectively. These three KPIs are:
  • The maximum end-to-end latency L is the maximum time elapsed between data generation by a sensor node and its delivery to the sink. To compute this value within the framework defined in Section 4.2, we make an additional assumption:
    Assumption 4.
    The maximum number of message transmissions on a link, denoted as M a x T r a n s , a parameter of the MAC TSCH protocol, is dynamically tuned according to the value computed by M F a i r or M O p t .
    With Assumption 4, the maximum end-to-end latency [27] is obtained when the last slot assigned to the node considered has just elapsed and then only the last transmission of the message is successful. This gives:
    ( S l o t f r a m e S i z e 1 + U s e d S l o t s ) × S l o t D u r a t i o n ,
    where U s e d S l o t s is the number of slots used by the schedule for data gathering. Hence, the smallest maximum end-to-end latency that can be achieved is obtained by an optimal schedule, which uses M i n S i z e the minimum number of slots for data gathering and for a slotframe duration equal to this number of slots. This smallest maximum latency is equal to
    ( 2 * M i n S i z e 1 ) × S l o t D u r a t i o n .
  • The end-to-end reliability R provided by the network. It is evaluated by the ratio of the total number of user-data messages sent by the sensor nodes over the total number of user-data messages delivered to the sink.
  • Network lifetime T is defined as the time the first node runs out of battery. Network lifetime can be expressed as:
    M i n N N o d e s I n i t i a l _ E n e r g y ( N ) × S F D u r a t i o n A v e r a g e _ E n e r g y _ C o n s u m p t i o n ( N ) ,
    where I n i t i a l _ E n e r g y ( N ) denotes the initial energy of node N, A v e r a g e _ E n e r g y _ C o n s u m p t i o n ( N ) is the average energy consumption of N per slotframe, and S F D u r a t i o n is the slotframe duration.
To evaluate the network lifetime, defined as the time up to first battery depletion of the busiest node, we use the parameters whose values are given in Table 2.
To summarize, the IoT network has to guarantee that at least R percent of the messages generated by sensor nodes are delivered to the sink with a latency L , whereas the network lifetime is at least equal to T.

4.4. Generalization of the Theoretical Bound on the Maximum Latency

We now compute a theoretical bound on the maximum latency when T X C e l l f ( N i ) < M a x T r a n s , where T X C e l l f ( N i ) denotes the number of TX cells assigned to flow f on node N i .
Let f be the flow whose message m has the maximum end-to-end latency. Let N k be the source node of f which is k hops away from the sink. Flow f visits successively N k , N k 1 , N 1 and then the sink. We adopt an additional assumption: on any visited node, m is never delayed by another flow. The worst case occurs when on N k message m is generated just after the last slot assigned to N k . Hence, N k has to wait for the next slotframe to transmit m. In addition, on any node, only the last transmission (i.e., the M a x T r a n s t h transmission) is received in the worst case; the previous ones are lost. According to the framework defined in Section 4.2, when any node N i receives a message of flow f in a slotframe, it has T X C e l l f ( N i ) cells to transmit it to its parent in the current slotframe. In each slotframe, any node N i has g T X C e l l g ( N i ) opportunities to transmit a message to its parent, where g is a flow visiting N i . Hence, we get the following formula:
M a x L a t e n c y ( f ) M a x T r a n s T X C e l l f ( N k ) + h = 1 k 1 M a x T r a n s T X C e l l f ( N h ) g T X C e l l g ( N h ) * S l o t f r a m e S i z e + S l o t U s e d * S l o t D u r a t i o n .
If only D e p t h , the routing tree depth, and M i n T X C e l l , the minimum number of TX cells per pair (sensor node, flow), are known, the bound becomes, taking into account that each sensor node generates its own flow:
M a x L a t e n c y M a x T r a n s M i n T X C e l l + h = 1 k 1 M a x T r a n s M i n T X C e l l M i n T X C e l l * ( D e p t h h + 1 ) * S l o t f r a m e S i z e + S l o t U s e d * S l o t D u r a t i o n .
Notice that Equations (9) and (10) generalize Equation (6), which is valid only when M i n T X C e l l M a x T r a n s .

5. Performance Results for a Toy Example

We first consider a toy example of a wireless TSCH network comprising a sink and seven sensor nodes. Each sensor node generates an application message of 27 bytes every 10 s. The slot duration is assumed to be 7.25 ms. The routing tree is depicted in Figure 3, where node A denotes the sink. The value associated with each link j gives P j the probability of successful receipt of a single transmission over that link. We notice that links have heterogeneous qualities, ranging from 0.5 to 0.9 . For each of the seven flows generated by a sensor node, we compute the maximum number of transmissions per link for any message of this flow. All the flows, except that generated by B, are multi-hop, which is six flows. In this example, we assume that Assumption 4 is met: M a x T r a n s is dynamically tuned according to the value computed by M F a i r or M o p t .

5.1. Number of Transmissions and End-To-End Reliability

Figure 4 depicts the total maximum number of transmissions for a message of a flow, depending on the hop number, when the targeted end-to-end reliability ranges from 0.9 to 0.9999. In this figure, the one-hop flow corresponds to the flow originating from B, whereas the two-hop flows correspond to the flows originating from C and E, the three-hop flows are those generated by D and F and the four-hop flows are those originating from G and H.
Table 3, Table 4, Table 5, Table 6 and Table 7 compare the values obtained with M F a i r and M O p t , when the targeted end-to-end reliability ranges from 0.9 to 0.99999 . These tables show that, whatever the targeted reliability, and as expected:
  • The maximum number of transmissions per message over a link increases, when the link quality decreases.
  • Both methods provide the same maximum number of transmissions for all single-hop flows.
  • The total number of transmissions per message of any given multi-hop flow obtained by M O p t is always less than or equal to that obtained by M F a i r . For instance, for R = 0.9 (see Table 3), we observe a gain on the total number of transmissions per message and per flow, which is equal to 1 for the 2-hop flows (i.e., flows originating from C and E), and for the 3-hop flow originating from D. This gain becomes 2 for the 4-hop flow originating from G and 3 for the 4-hop flow originating from H. To summarize the results obtained for the six multi-hop flows considered, we observe five improvements for R = 0.9 , four improvements for R = 0.99 , two improvements for R = 0.999 , four improvements for R = 0.9999 and five improvements for R = 0.99999 , leading to a total of 20 improvements over the 30 cases tested.
  • Even if the total number of transmissions is the same for both methods, the distribution over the links may differ as exemplified in Table 7 by the flow originating from D, where M O p t gets an end-to-end reliability of 0.99922 , whereas M F a i r gets a slightly less value 0.99921 . Hence, M O p t provides the smallest total number of transmissions per flow and, if equal with M F a i r , M O p t provides the highest end-to-end reliability.
  • With M F a i r , two links having the same link quality always have the same maximum number of transmissions for the same flow, as exemplified in all tables by the flow originating from H, where the links H D and C B have the same quality. However, this is not always the case with M O p t ; see, for instance, this flow in Table 4, where the maximum transmission number for link H D is 9, whereas it is 8 for link C B . Since the nodes close to the sink usually have a larger load, decreasing their load improves the network performances. Notice, however, that the maximum number of transmissions on a given link depends on the flow. For instance, the maximum number of transmissions on Link B A is equal to 11 for all flows, except the flow originating at B, where it is 10, for a targeted R = 0.99999 .
  • The number of iterations of M O p t never exceeds h + 1 in all the cases evaluated.

5.2. Load-Based Scheduling

Let us see how these transmissions are scheduled, assuming a per-flow approach and more precisely the selection of the Load-based scheduler, which schedules first the flow originating from the most loaded node. We recall that the load of a node is computed as the number of cells needed to transmit its own flows, plus the number of cells needed to receive and transmit the flows originating from its descendants. Figure 5 and Figure 6 depict the Load-based schedule of the seven flows generated by sensor nodes with M F a i r and M O p t , respectively, assuming a targeted end-to-end reliability of 0.9 . In both cases, the Load-based scheduler schedules the flows in the same order, starting with the most loaded node B, the scheduling order is B , C , D , E , H , F , G , although L o a d ( B ) = 52 cells with M F a i r and only 46 with M O p t . The two resulting schedules are optimal in terms of slots needed because node B, the most loaded node, is kept busy in all slots of both schedules. However, M F a i r requires exactly 52 slots to schedule the 72 transmissions, (see Figure 5), whereas M O p t requires only 46 slots to schedule the 64 transmissions, (see Figure 6). In this simple configuration with seven flows, M O p t allows for saving eight transmissions, which represents an improvement of 11 % in the number of transmissions to schedule. In addition, M O p t allows for saving six slots, which reduces by 11.5 % the number of slots used.
In the Load-based schedule obtained with M F a i r and depicted in Figure 5, we have M i n S i z e = 52 slots. Hence, for a slot duration of 7.25 ms, the smallest maximum latency that can be achieved with M F a i r is equal to ( 52 + 51 ) * 7.25 = 0.7465 s. The average energy consumption per slotframe of node B, the greatest loaded node, is equal to: ( 22 × T X C h a r g e + 30 × R X C h a r g e ) / S F D u r a t i o n , where S F D u r a t i o n denotes the slotframe duration. With an initial energy of 2821.5 mAh, the default slotframe size of 101 slots and a slot duration of 7.25 ms, this node will have a lifetime of 39 days and a maximum latency of ( 101 1 + 52 ) * 7.25 * 10 3 = 1.102 s. To meet a lifetime of one year, the slotframe size should be greater than or equal to 933 slots, with a maximum latency of ( 933 1 + 52 ) * 7.25 * 10 3 = 7.0905 s.
With the Load-based schedule obtained with M O p t and depicted in Figure 6, we have M i n S i z e = 46 slots. Hence, for a slot duration of 7.25 ms, the smallest maximum latency that can be achieved with M O p t is equal to ( 46 + 45 ) * 7.25 = 0.65975 s, which represents an improvement of 13.67 % . It becomes 1.0585 s for the default slotframe size of 101 slots. The average energy consumption of B per slotframe becomes ( 20 × T X C h a r g e + 26 × R X C h a r g e ) / S F D u r a t i o n , leading to a network lifetime of 44 days for the default slotframe size of 101 slots and a maximum end-to-end latency of ( 101 1 + 46 ) * 7.25 = 1.0585 s. To meet a lifetime of one year, the slotframe size should be greater than or equal to 830 slots, with a maximum latency of ( 830 1 + 46 ) * 7.25 * 10 3 = 6.34375 s, which represents an improvement of 12.38 % with regard to M F a i r . For a slotframe size of 933 slots, M O p t would provide a maximum latency of 7.0905 s, a decrease of 11.77% with regard to M F a i r and a network lifetime of 410 days instead of 365 for M F a i r , an increase of 12.40%.
Table 8 points out the trade-off between the maximum end-to-end latency and network lifetime by listing the results obtained by M F a i r and M O p t for different slotframe sizes: 52, 101 and 933 slots. To increase the network lifetime by increasing the slotframe size, provided that the application still generates the same number of messages per slotframe, leads to an increase in maximum latency. M O p t provides a shorter maximum end-to-end latency because of a smaller schedule size (i.e., smaller number of slots used).

6. Performance Results of a TSCH Network with 50 Nodes

For the performance evaluation of a TSCH network with 50 nodes, we use the 6TiSCH simulator [28], which has been designed for a fast prototyping. In [8], network performances are evaluated on two specific applications running on a same network. In this particular configuration, randomly deployed sensors are only in charge of generating messages that they forward to a close relay, whereas relays are deployed according to a triangular grid. Since a more generic configuration is representative of much more applications, we focus on a generic data gathering application running on random network topologies.

6.1. Simulation Parameters

The network topology is a random topology such that any mote has at least three neighbors (i.e., three motes with which P D R 0.5 ). For each wireless link, the PDR value is computed according to the Pister–Hack model [29]. The 6TiSCH protocol stack is used with RPL as the routing protocol with the E T X metric and the Load-based scheduler as the scheduling function. RPL, M F a i r and M O p t use the PDR values computed for the wireless links considered. For each wireless link i, RPL deduces E T X ( i ) from P D R ( i ) . The simulation parameters used to evaluate the KPIs are those given in Table 9. The deepest routing tree observed in the simulations is 7-hop deep, the shallowest is 4-hop deep, with a median of 5-hop.
Notice that, in the simulations done with the 6TiSCH stack, M a x T r a n s , the maximum number of transmissions of any message on any link is fixed, as in the standardized MAC TSCH protocol. In the 6TiSCH simulations, its value is set to 6. If after six transmissions the acknowledgment is not received, the message is discarded. This behavior has a strong impact on the latency. Since Assumption 4 is not met, the theoretical bound for the maximum latency given in Equation (6) is no longer valid, we use the new theoretical bound given in Section 4.4.
A legitimate question is why a schedule provides a number of Transmission cells (TX) for any given flow greater than M a x T r a n s , since a message that has not been acknowledged after M a x T r a n s transmissions is discarded. The justification is provided by the decrease in the average end-to-end latency and the maximum end-to-end latency as we will see in Section 6.3. This decrease is due to a greater number of opportunities to transmit.

6.2. End-To-End Delivery Rate

Simulation results about the average end-to-end delivery rate are depicted in Figure 7.
As expected, the end-to-end reliability is better with M O p t than with M F a i r because M O p t maximizes the end-to-end reliability provided for a minimum total number of transmissions per message.

6.3. End-To-End Latency

Simulation results about the average end-to-end latency are depicted in Figure 8.
As M F a i r tends to compute a greater number of transmissions than M O p t , the schedule for M F a i r includes a greater number of cells in a slotframe whose size is kept identical for M O p t and M F a i r . The more cells, the more chance to send or forward. Since an application packet is generated at a random time point in the slotframe, the more transmission (TX) cells the node has, and the more chance to send the packet immediately. This is why the average end-to-end latency is shorter with M F a i r .
Figure 9 shows the percentage of messages delivered in a single slotframe. It is smaller with M O p t than with M F a i r due to the greater number of TX cells granted by M F a i r .
Figure 10 depicts the cumulative distribution function of the end-to-end latency for M F a i r and M O p t . With M F a i r , 98% of messages reach the sink in one slotframe that is with an end-to-end latency ≤7 s. The gap between M F a i r and M O p t is smaller than 1 % .
The maximum end-to-end latency obtained by simulation with M F a i r and M O p t is illustrated in Figure 11. Since M F a i r allocates more TX cells and the slotframe size is kept identical for M F a i r and M O p t , M F a i r provides a shorter maximum end-to-end latency. For targeted reliabilities ≥0.999, M F a i r and M O p t provide very close maximum latencies.
Figure 12 depicts the schedule size expressed as the number of slots used in the schedule of M F a i r and M O p t . Unsurprisingly, the schedule size increases with the targeted end-to-end reliability, due to a great number of transmissions on each link to reach the targeted end-to-end reliability. Whatever the targeted reliability, the schedule size is always shorter with M O p t than with M F a i r , as expected.
Table 10 compares the theoretical bound for the maximum end-to-end latency obtained by Equation (10) with the simulation results, for different values of the targeted end-to-end reliability and a given random topology. The maximum end-to-end latency obtained by simulation decreases when the targeted end-to-end reliability increases: a greater number of TX cells assigned to nodes give them more opportunities to transmit in a slotframe. With the theoretical bound, the decrease is obtained when the number of TX Cells assigned to a node increases to reach the targeted end-to-end reliability. Whatever the targeted reliability, the bound given by Equation (10) and the simulation results are not close. The theoretical bound could be refined to take more information into account.
Table 10 also provides the end-to-end reliability. As expected, M O p t provides an end-to-end reliability better than M F a i r . However, for a targeted end-to-end reliability ≥0.999, M F a i r fails to achieve the requested reliability, when M a x T r a n s is left fixed to 6 instead of being dynamically tuned according to the values computed by M F a i r or M O p t .

6.4. Duty Cycle

With regard to network lifetime and energy consumption, we consider the busiest node excluding the root (sink) node, since it is supposed to be mains-powered. This busiest node determines the network lifetime. The duty cycle on this busiest node is computed as:
D u t y _ C y c l e = n u m b e r _ o f _ s l o t s _ u s e d _ o n _ t h e _ b u s i e s t _ n o d e s c h e d u l e _ s i z e _ i n _ s l o t s .
Note that Equation (11) is meant for the following comparison, which does not represent actual radio duty cycle. In TSCH, a device does not turn on its radio all the time even during an active slot.
Figure 13 depicts the number of cells assigned to the busiest node in one simulation run for each targeted end-to-end reliability.
Table 11 gives the duty cycle on the busiest node with M F a i r and M O p t . This gives an insight on network lifetime which is determined by the lifetime of the busiest node.
As a consequence, for high targeted reliabilities (i.e., ≥0.999), M F a i r and M o p t give close end-to-end latencies at a greater energy cost for M F a i r .

6.5. Impact of M a x T r a n s , a TSCH Parameter

We now study the impact of M a x T r a n s the maximum number of transmissions of a message in TSCH, whose default value is 6. We set M a x T r a n s to the value of 20, which is greater than the maximum number of transmissions on each link computed by M F a i r or M O p t . We run the same simulations as previously and evaluate the impact on the end-to-end reliability and the end-to-end latency.
As expected, the end-to-end reliability depicted in Figure 14 is increased both with M O p t and M F a i r , when the maximum number of transmissions per message in TSCH is set to 20. It is very close to 100%, for all the values of the targeted reliability tested. Notice, however, that increasing M a x T r a n s may lead to a network overload resulting in violations of the maximum acceptable latency L. To avoid that, messages whose lifetime is greater than or equal to L should not be transmitted and should be discarded.
The consequence of increasing M a x T r a n s to 20 is an increase in the average latency for both M F a i r and M O p t , as shown in Figure 15. The shortest average latency is still provided by M F a i r . However, the gap decreases for high targeted end-to-end reliabilities (i.e., ≥0.999). Using a value of M a x T r a n s greater than the value required by M F a i r or M O p t on a given link may result in an increase in end-to-end latency and energy consumption for a very strongly limited gain in end-to-end reliability.

7. Comparison with Kausa

Since K a u s a [8] is a well-known centralized scheduler that takes into account the unreliability of wireless links and adopts a per-flow approach to meet KPIs, we now compare M O p t to K a u s a when run in a 6TiSCH network.

7.1. Our K a u s a Implementation in the 6TiSCH Simulator

In the 6TiSCH simulator, as in any 6TiSCH implementation compliant with the standard, routing is done by the RPL protocol with the ETX metric. We simulated K a u s a on the 6TiSCH simulator and ran it with the generic configuration defined in Section 6.1. We recall that the value of M a x T r a n s is fixed to 6. In these conditions, the main differences between Kausa and our approach are:
  • K a u s a selects first the flow requesting the highest end-to-end reliability, then the shortest end-to-end latency and finally the flow originating from the farthest node of the sink. Our approach selects the flow originating from the most loaded sensor node (i.e., the sensor node needing the largest number of Tx+Rx cells).
  • For any flow f, K a u s a starts by assigning cells to the most loaded node visited by f. Then, K a u s a goes backward to the source of f. Finally, K a u s a goes upward from the most loaded node to the sink. In our approach, cells are assigned to nodes visited by f in a cascading way from the source of f up to the sink. It follows that our approach is easier to implement.
  • For any flow f, K a u s a minimizes the number of retransmissions on the most loaded node, whereas we minimize the total number of retransmissions on the path of f.
Notice that we implement an optimization of K a u s a for 6TiSCH, enabling any node to use its next Tx cell to transmit any message to its parent, as done in our approach. This is not true in the published version of K a u s a , where any node N is not allowed to use cells assigned to a flow f for another flow f visiting N.

7.2. End-To-End Latency

Figure 16 depicts the average end-to-end latency obtained by M O p t and K a u s a . They both provide close values. However, M O p t provides a much smaller variance of the average end-to-end latency than K a u s a , making it more predictable, even for high targeted reliabilities.
The same conclusion applies to the maximum end-to-end latency depicted in Figure 17. The predictability of performance is a property sought by industrial applications.

7.3. End-To-End Reliability

With regard to the end-to-end PDR, we observe in Figure 18 that both M O p t and K a u s a ensure a PDR greater than the targeted one when it belongs to the interval [0.9, 0.999]. For greater values, the PDR is not met because the value of M a x T r a n s = 6 is too small to reach the targeted end-to-end reliability. In addition, M O p t tends to provide a greater median value than K a u s a .

7.4. Duty Cycle

Since the energy consumption can be deduced from the duty cycle of the busiest node, we compare the number of cells scheduled at the busiest node by K a u s a and M O p t , as depicted in Figure 19. Both provide close values for a targeted end-to-end reliability less than or equal to 0.999. For a higher reliability, K a u s a provides a smaller number of cells due to a number of transmissions scheduled per message less than or equal to M a x T r a n s , whereas M O p t may schedule a larger number, explaining this result.

7.5. Schedule Size

The schedule size, illustrated in Figure 20, is a little greater with M O p t than with K a u s a . As a consequence, a node has fewer opportunities to transmit with K a u s a than with M O p t , leading to a higher and less predictable end-to-end latency.

8. Conclusions

TSCH is a very promising technology for the IoT. It is now necessary to evaluate the performances it can provide to IoT applications. These performances are evaluated by means of three KPIs: maximum end-to-end latency L, end-to-end reliability R and network lifetime T. This IoT network has to guarantee that at least R % of messages generated by sensors are delivered to the sink with a latency less than or equal to L, while the network lifetime is at least T. In this paper, we have proposed two methods M F a i r and M O p t to achieve a targeted end-to-end reliability taking into account the unreliability of wireless links. The trade-offs between end-to-end latency, network lifetime and end-to-end reliability have been pointed out. In addition, we have shown that minimizing the total number of transmissions of a message to reach the sink with M O p t saves 12 % of network lifetime in a small network of eight nodes, assuming a Load-based schedule of flows. As expected, M O p t provides a better end-to-end reliability and a longer network lifetime than M F a i r . However, the average end-to-end latency provided by M F a i r is smaller. These results scale up to a 50-node network representative of real deployments, as shown by simulation results obtained with the 6TiSCH simulator. However, using a fixed maximum number of transmissions M a x T r a n s equal to the default value of the TSCH protocol, instead of using the value computed by M F a i r or M O p t for the link and the flow considered, may lead to a violation of the targeted end-to-end reliability: messages that have not been acknowledged after M a x T r a n s transmissions are discarded. Compared to K a u s a , a KPI-aware, state-of-the-art scheduler, our approach is simpler to implement and ensures more predictable end-to-end performances, which is an essential property for industrial applications.

Author Contributions

Y.T. implemented the MFair and MOpt algorithms, as well as the Load-based scheduler and Kausa in the 6TiSCH simulator. He designed the performance evaluation and performed it for many random network configurations and four values of targeted end-to-end reliability. He drew the graphs corresponding to simulation results and analyzed them. P.M. designed MOpt and compared MFair and MOpt on the toy example. She also computed the theoretical values of the end-to-end latency and wrote the paper.

Funding

This research received no external funding.

Acknowledgments

The authors acknowledge Keoma Brun-Laguna and Thomas Watteyne for their valuable feedback on this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Delaney, D.; O’Hare, G. A framework to implement IoT network performance modelling techniques for network solution selection. Sensors 2016, 16, 2038. [Google Scholar] [CrossRef] [PubMed]
  2. Tao, M.; Ota, K. Locating compromised data sources in IoT-enabled smart cities: A great-alternative-region-based approach. IEEE Trans. Ind. Inf. 2018, 14, 2579–2587. [Google Scholar] [CrossRef]
  3. Tao, M.; Wei, W.; Huang, S. Location-based trustworthy services recommendation in cooperative- communication-enabled internet of vehicles. J. Netw. Comput. Appl. 2019, 126, 1–11. [Google Scholar] [CrossRef]
  4. Praveen Kumar, M.; Santhoshkumar, S.; Gowdhaman, T.; Syed Shajahaan, S. A survey on IoT performances in big data. Int. J. Comput. Sci. Mob. Comput. IJCSMC 2017, 6, 26–34. [Google Scholar]
  5. Srinidhi, N.; Dilip Kumar, S.; Venugopal, K. Network optimizations in the Internet of Things: A review. Eng. Sci. Technol. Int. J. 2019, 22, 1–21. [Google Scholar] [CrossRef]
  6. IEEE Standards Association. IEEE Standard for Local and Metropolitan Area Networks—Part 15.4: Low-Rate Wireless Personal Area Networks (LR-WPANs); IEEE Std 802.15.4-2011 (Revision of IEEE Std 802.15.4-2006); IEEE: Piscataway, NJ, USA, 2011. [Google Scholar] [CrossRef]
  7. IEEE Standards Association. IEEE Standard for Local and Metropolitan Area Networks—Part 15.4: Low-Rate Wireless Personal Area Networks (LR-WPANs)—Amendment 1: MAC Sublayer; IEEE Std 802.15.4e-2012 (Amendment to IEEE Std 802.15.4-2011); IEEE: Piscataway, NJ, USA, 2012. [Google Scholar]
  8. Gaillard, G.; Barthel, D.; Theoleyre, F.; Valois, F. Kausa: KPI-aware scheduling algorithm for multi-flow in multi-hop IoT networks. In Ad-Hoc, Mobile, and Wireless Networks; Mitton, N., Loscri, V., Mouradian, A., Eds.; Springer International Publishing: Lille, France, 2016; pp. 47–61. [Google Scholar]
  9. Kirubasri, G.; Uma Maheswari, N. A study on hardware and software link quality metrics for wireless multimedia sensor networks. Int. J. Adv. Netw. Appl. 2016, 8, 3103–3109. [Google Scholar]
  10. Baccour, N. Radio link quality estimation in wireless sensor networks: A survey. ACM Trans. Sens. Netw. (TOSN) 2012, 8, 35. [Google Scholar] [CrossRef]
  11. De Couto, D.; Aguayo, D.; Bicket, J.; Morris, R. A high-throughput path metric for multi-hop wireless routing. In Proceedings of the 9th Annual International Conference on Mobile Computing and Networking, San Diego, CA, USA, 14–19 September 2003; Volume 11, pp. 134–146. [Google Scholar] [CrossRef]
  12. Baccour, N.; Koubâa, A.; Youssef, H.; Ben Jamâa, M.; do Rosário, D.; Alves, M.; Becker, L.B. F-LQE: A fuzzy link quality estimator for wireless sensor networks. In Wireless Sensor Networks; Silva, J.S., Krishnamachari, B., Boavida, F., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 240–255. [Google Scholar]
  13. Pöttner, W.B.; Seidel, H.; Brown, J.; Roedig, U.; Wolf, L. Constructing schedules for time-critical data delivery in wireless sensor networks. ACM Trans. Sen. Netw. 2014, 10, 44. [Google Scholar] [CrossRef]
  14. Vlavianos, A.; Law, L.K.; Broustis, I.; Krishnamurthy, S.V.; Faloutsos, M. Assessing link quality in IEEE 802.11 wireless networks: Which is the right metric? In Proceedings of the 2008 IEEE 19th International Symposium on Personal, Indoor and Mobile Radio Communications, Cannes, France, 15–18 September 2008; IEEE: Piscataway, NJ, USA, 2008. [Google Scholar]
  15. Cerar, G.; Mohorcic, M.; Gale, T.; Fortuna, C. Link Quality Estimation using Machine Learning. arXiv 2019, arXiv:1812.08856. [Google Scholar]
  16. Sun, W.; Lu, W.; Li, Q.; Chen, L.; Mu, D.; Yuan, X. WNN-LQE: Wavelet-neural-network-based link quality wstimation for smart grid WSNs. IEEE Access 2017, 5, 12788–12797. [Google Scholar] [CrossRef]
  17. Rekik, S.; Baccour, N.; Jmaiel, M.; Drira, K. Low-Power link quality estimation in smart grid environments. In Proceedings of the 2015 International Wireless Communications and Mobile Computing Conference (IWCMC), Dubrovnik, Croatia, 24–28 August 2015; pp. 1211–1216. [Google Scholar] [CrossRef]
  18. Guo, Z.; Wang, Q.; Li, M.; He, J. Fuzzy logic based multidimensional link quality estimation for multi-hop wireless sensor networks. IEEE Sens. J. 2013, 13, 3605–3615. [Google Scholar] [CrossRef]
  19. Nguyen, D.; Minet, P. Quality of service routing in a MANET with OLSR. J. Univ. Comput. Sci. 2007, 13, 56–86. [Google Scholar]
  20. Marinca, D.; Minet, P. On-line learning and prediction of link quality in wireless sensor networks. In Proceedings of the 2014 IEEE Global Communications Conference GLOBECOM 2014, Austin, TX, USA, 8–12 December 2014; IEEE: Piscataway, NJ, USA, 2014. [Google Scholar]
  21. Mahfoudh, S.; Minet, P.; Amdouni, I. Energy efficient routing and node activity scheduling in the OCARI wireless sensor network. Future Internet 2010, 2, 308–340. [Google Scholar] [CrossRef]
  22. Gaillard, G.; Barthel, D.; Theoleyre, F.; Valois, F. Enabling Flow-Level Reliability on FTDMA Schedules with Efficient Hop-by-Hop Over-Provisioning. Available online: https://hal.inria.fr/hal-01278077v2 (accessed on 13 September 2019).
  23. Palattella, M.R.; Accettura, N.; Dohler, M.; Grieco, L.A.; Boggia, G. Traffic Aware Scheduling Algorithm for reliable low-power multi-hop IEEE 802.15.4e networks. In Proceedings of the 2012 IEEE 23rd International Symposium on Personal, Indoor and Mobile Radio Communications—(PIMRC), Sydney, NSW, Australia, 9–12 September 2012; pp. 327–332. [Google Scholar] [CrossRef]
  24. Audéoud, H.; Heusse, M. Quick and efficient link quality estimation in wireless sensors networks. In Proceedings of the 2018 14th Annual Conference on Wireless On-demand Network Systems and Services (WONS), Isola, France, 6–18 February 2018; pp. 87–90. [Google Scholar] [CrossRef]
  25. Brun-Laguna, K.; Henrique Gomes, P.; Minet, P.; Watteyne, T. Moving beyond testbeds? Lessons (we) learned about connectivity. IEEE Pervasive Comput. 2018, 17, 15–27. [Google Scholar] [CrossRef]
  26. Brun-Laguna, K.; Minet, P.; Tanaka, Y. Optimized scheduling algorithm for time-critical industrial IoT. In Proceedings of the IEEE Global Communications Conference (Globecom), Waikoloa, HI, USA, 9–13 December 2019. [Google Scholar]
  27. Khoufi, I.; Minet, P.; Rmili, B. Scheduling transmissions with latency constraints in an IEEE 802.15.4e TSCH network. In Proceedings of the VTC 2017—IEEE 86th Vehicular Technology Conference, Toronto, ON, Canada, 24–27 September 2017. [Google Scholar]
  28. Municio, E.; Daneels, G.; Vicinic, M.; Latre, S.; Famaey, J.; Tanaka, Y.; Brun-Laguna, K.; Muraoka, K.; Vilajosana, X.; Watteyne, T. Simulating 6TiSCH networks. Trans. Emerging Telecommun. Technol. 2018, 30, e3494. [Google Scholar] [CrossRef]
  29. Le, H.P.; John, M.; Pister, K. Energy-Aware Routing in Wireless Sensor Networks with Adaptive Energy-Slope Control; Technical Report EE290Q-2 Spring; University of California: Berkeley, CA, USA, 2009. [Google Scholar]
Figure 1. Maximum number of transmissions over a link visited by a 4-hop flow, computed by M F a i r .
Figure 1. Maximum number of transmissions over a link visited by a 4-hop flow, computed by M F a i r .
Sensors 19 03970 g001
Figure 2. Expected number of transmissions versus M F a i r Maximum number of transmissions over a link.
Figure 2. Expected number of transmissions versus M F a i r Maximum number of transmissions over a link.
Sensors 19 03970 g002
Figure 3. A routing tree with the PDR of each link used by flows.
Figure 3. A routing tree with the PDR of each link used by flows.
Sensors 19 03970 g003
Figure 4. Total number of transmissions for a flow message.
Figure 4. Total number of transmissions for a flow message.
Sensors 19 03970 g004
Figure 5. Load-based schedule with M F a i r .
Figure 5. Load-based schedule with M F a i r .
Sensors 19 03970 g005
Figure 6. Load-based schedule with M O p t .
Figure 6. Load-based schedule with M O p t .
Sensors 19 03970 g006
Figure 7. End-to-end delivery rate obtained with M F a i r and M O p t .
Figure 7. End-to-end delivery rate obtained with M F a i r and M O p t .
Sensors 19 03970 g007
Figure 8. Average end-to-end latency obtained with M F a i r and M O p t .
Figure 8. Average end-to-end latency obtained with M F a i r and M O p t .
Sensors 19 03970 g008
Figure 9. Percentage of packets delivered in one slotframe with M F a i r and M O p t .
Figure 9. Percentage of packets delivered in one slotframe with M F a i r and M O p t .
Sensors 19 03970 g009
Figure 10. Cumulative distribution function of the end-to-end latency with M F a i r and M O p t .
Figure 10. Cumulative distribution function of the end-to-end latency with M F a i r and M O p t .
Sensors 19 03970 g010
Figure 11. Maximum end-to-end latency obtained by simulation with M F a i r and M O p t .
Figure 11. Maximum end-to-end latency obtained by simulation with M F a i r and M O p t .
Sensors 19 03970 g011
Figure 12. Maximum slot used by M F a i r and M O p t .
Figure 12. Maximum slot used by M F a i r and M O p t .
Sensors 19 03970 g012
Figure 13. Number of cells assigned by M F a i r and M O p t on the busiest node.
Figure 13. Number of cells assigned by M F a i r and M O p t on the busiest node.
Sensors 19 03970 g013
Figure 14. End-to-end delivery rate obtained with M F a i r and M O p t for M a x T r a n s = 20 transmissions.
Figure 14. End-to-end delivery rate obtained with M F a i r and M O p t for M a x T r a n s = 20 transmissions.
Sensors 19 03970 g014
Figure 15. Average end-to-end latency obtained with M F a i r and M O p t for MaxTrans = 20 transmissions.
Figure 15. Average end-to-end latency obtained with M F a i r and M O p t for MaxTrans = 20 transmissions.
Sensors 19 03970 g015
Figure 16. Average end-to-end latency obtained with M O p t and K a u s a .
Figure 16. Average end-to-end latency obtained with M O p t and K a u s a .
Sensors 19 03970 g016
Figure 17. Maximum end-to-end latency obtained with M O p t and K a u s a .
Figure 17. Maximum end-to-end latency obtained with M O p t and K a u s a .
Sensors 19 03970 g017
Figure 18. End-to-end delivery rate obtained with M O p t and K a u s a .
Figure 18. End-to-end delivery rate obtained with M O p t and K a u s a .
Sensors 19 03970 g018
Figure 19. Number of cells used at the busiest node with M O p t and K a u s a .
Figure 19. Number of cells used at the busiest node with M O p t and K a u s a .
Sensors 19 03970 g019
Figure 20. Schedule size obtained with M O p t and K a u s a .
Figure 20. Schedule size obtained with M O p t and K a u s a .
Sensors 19 03970 g020
Table 1. Main notations.
Table 1. Main notations.
NameMeaning
Rthe targeted end-to-end reliability
Lthe targeted end-to-end latency
Tthe targeted network lifetime
jthe link from node N j to its parent in the routing tree
M j the maximum number of transmissions of any message transmitted on link j
P j the probability of successful acknowledgment receipt after a single message transmission on link j
R j the probability of successful acknowledgment receipt on link j after a maximum number M j of transmissions
Table 2. Parameters used for network lifetime computation.
Table 2. Parameters used for network lifetime computation.
ParameterValue
Initial energy of node powered by 2 Energizers L-91 AA batteries2821.5 mAh
Transmit a data frame & receive its acknowledgment54.5 μ C
Receive a data frame & transmit its acknowledgment32.6 μ C
IdleListen6.4 μ C
Sleep0 μ C
Table 3. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.9 on seven flows.
Table 3. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.9 on seven flows.
Maximum Number of Transmissions for Targeted R = 0.9Total Number of Transmissions per msgEnd-To-End Reliability
FlowHopP = 0.7P = 0.5P = 0.8P = 0.6P = 0.9
FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.
B122 220.910.91
C23354 870.94250.91218
E233 43 760.94800.9107
D3335532 11100.93500.90489
F333 44 10100.922490.92249
G4436533 2215130.958900.92570
H4436533 19160.953450.90583
Table 4. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.99 on seven flows.
Table 4. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.99 on seven flows.
Maximum Number of Transmissions for Targeted R = 0.99Total Number of Transmissions per msgEnd-To-End Reliability
FlowHopP = 0.7P = 0.5P = 0.8P = 0.6P = 0.9
FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.
B144 440.99190.9919
C25588 13130.9936730.993673
E255 66 11110.993480.99348
D3559844 18170.994020.99208
F3555 76 17160.99350.99106
G4559844 3321200.993030.99109
H45599 833 27260.992080.99014
Table 5. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.999 on seven flows.
Table 5. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.999 on seven flows.
Maximum Number of Transmissions for Targeted R = 0.99Total Number of Transmissions per msgEnd-To-End Reliability
FlowHopP = 0.7P = 0.5P = 0.8P = 0.6P = 0.9
FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.
B166 660.999270.99927
C2771111 18180.999290.99929
E277 98 16150.999510.99912
D377121156 24240.999210.999229
F377 99 23230.999300.99930
G478121166 4429280.999370.99922
H477121266 37370.9992290.999229
Table 6. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.9999 on seven flows.
Table 6. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.9999 on seven flows.
Maximum Number of Transmissions for Targeted R = 0.99Total Number of Transmissions per msgEnd-To-End Reliability
FlowHopP = 0.7P = 0.5P = 0.8P = 0.6P = 0.9
FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.
B188 880.9999340.999934
C2991515 24240.99994980.9999498
E299 1111 20200.999938370.99993837
D399151477 31300.9999370.999918
F399 1211 30290.999943860.999906
G499161577 5537360.9999420.999927
H499161577 48460.9999370.999906
Table 7. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.99999 on seven flows.
Table 7. Comparison of M F a i r and M O p t for a targeted end-to-end reliability R = 0.99999 on seven flows.
Maximum Number of Transmissions for Targeted R = 0.99Total Number of Transmissions per msgEnd-To-End Reliability
FlowHopP = 0.7P = 0.5P = 0.8P = 0.6P = 0.9
FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.FairOpt.
B11010 10100.99999410.9999941
C211111817 29280.999994410.9999906
E21111 1413 25240.999995540.99999152
D31111191888 38370.999993760.99999185
F31111 1414 36360.999993770.99999377
G41111191899 6645440.99999480.9999929
H41111191899 58560.99999390.9999909
Table 8. The trade-off of end-to-end latency versus network lifetime for M F a i r and M O p t .
Table 8. The trade-off of end-to-end latency versus network lifetime for M F a i r and M O p t .
MFair MOpt Relative Improvement
Slotframe ( slots ) Slotframe ( slots ) Slotframe ( slots )
521019335210193352101933
Max. latency (s)0.746751.1027.1340.703251.05857.09055.82%3.94%0.61%
Lifetime (days)20.3539.54365.2822.8744.42410.4110.98%12.35%12.35%
Table 9. Simulation parameters.
Table 9. Simulation parameters.
ParameterValue
Config.Number of nodes50
Number of Channels16
Topologyrandom
Link reliabilitycomputed
TSCHSlot duration10 ms
Slotframe size700 slots
Secure joindisabled
Keep-alive (L2)disabled
Transmit queue size10 packets
Max of retransmissions5 (i.e., max of 6 transmissions)
RoutingRPLwith ETX metric where DAO is disabled
a stable routing topologyafter 60 min
SchedulingScheduling functionLoad-based scheduler
Applicationis run & measures madeduring the next 60 min
packet generation interval on each sensor node[57 s, 63 s]
Simulation100 runs per pair (algorithm, targeted reliability)algorithm { M O p t , M F a i r }
target. reliability { 0.9 , 0.99 , 0.999 , 0.9999 }
Table 10. Comparison of the theoretical maximum end-to-end latency and the simulation results for M F a i r and M O p t , taken from particular simulations having the same routing topology.
Table 10. Comparison of the theoretical maximum end-to-end latency and the simulation results for M F a i r and M O p t , taken from particular simulations having the same routing topology.
Targeted End-To-End Reliability
0.90.990.9990.9999
Schedule size (slots) M F a i r 96154199249
M O p t 84134190246
relative improvement12.5%12.98%4.52%1.20%
Min TX cells per (node, flow) M F a i r 2223
M O p t 2223
Maximum end-to-end latency (s) M F a i r theory28.9629.5429.9916.49
simulation13.7913.939.647.70
M O p t theory28.8429.3429.9016.46
simulation19.9313.947.437.69
End-to-end reliability M F a i r 0.9979590.9989770.9989780.999659
M O p t 10.9982940.9993181
Table 11. Comparison of the duty cycle of the busiest node for M F a i r and M O p t .
Table 11. Comparison of the duty cycle of the busiest node for M F a i r and M O p t .
Targeted End-To-End Reliability
0.90.990.9990.9999
Schedule size (slots) M F a i r 96154199249
M O p t 84134190246
Cells assigned to busiest node M F a i r 77103127159
M O p t 6387122149
Duty cycle of busiest node M F a i r 11%14.71%18.14%22.71%
M O p t 9%12.42%17.42%21.28%
Relative gain18.18%15.53%3.93%6.29%

Share and Cite

MDPI and ACS Style

Minet, P.; Tanaka, Y. Optimal Number of Message Transmissions for Probabilistic Guarantee of Latency in the IoT. Sensors 2019, 19, 3970. https://doi.org/10.3390/s19183970

AMA Style

Minet P, Tanaka Y. Optimal Number of Message Transmissions for Probabilistic Guarantee of Latency in the IoT. Sensors. 2019; 19(18):3970. https://doi.org/10.3390/s19183970

Chicago/Turabian Style

Minet, Pascale, and Yasuyuki Tanaka. 2019. "Optimal Number of Message Transmissions for Probabilistic Guarantee of Latency in the IoT" Sensors 19, no. 18: 3970. https://doi.org/10.3390/s19183970

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop