Keywords

1 Introduction

The number of Internet of Things (IoT) devices is increasing rapidly with the increasing needs of smart cities, healthcare applications, autonomous systems, and smart vehicles [6, 7, 12, 40], causing the overload of communication channels and gateways [27]. This results in the Massive Access Problem (MAP) where high latency and queue lengths can lead to packet loss and deadline violations. In addition, congestion can lead to increased energy consumption at IoT devices and gateways due to repeated requests for access and increased processing times [2, 15], thus contributing to the worldwide increase in energy consumption for ICT [14].

Thus, substantial work over the last several years [3, 11, 28,29,30, 34, 35, 39, 46,47,48,49,50,51, 55, 56] attempts to solve MAP in various ways.

In this paper, we first briefly review methodologies and results with regard to reactive or proactive (predictive) solutions that can mitigate MAP. Then, we summarize two recent research avenues: Randomization of Transmission Times [36] and a novel traffic shaping policy – the Quasi-Deterministic-Transmission-Policy (QDTP) [18, 19, 25]. We illustrate the results these approaches offer via analytical techniques and trace driven simulations using a publicly available dataset of up to 6400 IoT devices [1] with different deadline constraints.

The remainder of this paper is organized as follows. Section 2 reviews recent research focusing on MAP. Section 3 summarizes two recent studies on MAP based on analytical and experimental results. Finally in Sect. 4 our main conclusions are presented.

2 Review of Prior Work on MAP

This section reviews the prior work on MAP in two categories as reactive solutions and predictive/proactive solutions. Early research addressed MAP by reducing congestion through adaptive Random Re-Routing (RRR) [20,21,22] which improves the QoS of a sensor network by dynamically changing packet routes when congestion is detected. In related work [42], an information theoretic technique selectively reduces the amount of traffic by increasing transmission efficiency, and improves the QoS in sensor networks.

More recent work has proposed solutions to MAP, assuming that IoT traffic is generated at random, and using approaches mostly based on Access Class Barring (ACB). In [35] ACB is enhanced by using Markov chains to model the status of preambles and to forecast active devices, while other work [30] developed the recursive ACB algorithm based on instantaneous detection of idle preambles. Recent work [28] also developed recursive ACB which adapts the probability that a device sends a preamble based on estimating the number of active IoT devices. The performance of ACB has been analyzed under different parameters for Machine-to-Machine (M2M) communications [55], and enhanced by Reinforcement Learning (RL) to select its parameter (i.e. barring rate) with respect to network conditions [56]. In [29] deep RL techniques are proposed to maximize the number of devices that successfully access the medium without collisions.

In [34] an access scheme for M2M communication that clusters machine-type devices according to their requirements and locations is used. The Non-Orthogonal Multiple Access (NOMA) based technique is presented for networks with a massive number of devices in [51]. Moreover, in order to address MAP, in [3] a collision detection based random access technique is developed, and in [49] a hybrid technique that combines slotted-Carrier-Sense Multiple Access with Collision Avoidance (CSMA/CA) and Time-Division Multiple Access (TDMA) schemes are suggested.

2.1 Proactive Solutions

IoT devices are quite simple and cannot easily be coordinated for timing or scheduling [24] based on distributed control [8, 10]. Thus recent experimental results [37, 38] show that machine learning techniques can be used to predict IoT traffic generated by individual devices, and other work [11, 39, 46,47,48, 50] designs proactive/predictive access schemes that determine the transmission times from IoT devices based on such predictions to mitigate the MAP.

Other work [31, 45, 48] has developed proactive access schemes for Machine-to-Human (M2H) or Human-to-Machine (H2M) traffic, while earlier research [43, 44] has focused on Human-to-Human (H2H) traffic. In [43] a schedule-based protocol an expert system is used to determine schedules that minimize delay and maximize channel utilization. In [44], forecasts of data rates of individual applications are used to schedule channel scheduling, and network load has also been balanced based on the forecast of the total load of all machine-type devices [31]. To mitigate the latency bottleneck due to the contention in optical or wireless networks [48], the proactive allocation of bandwidth to transmissions of packet bursts based on Artificial Neural Network (ANN) forecasts is studied, while in [45] the prediction of network throughput for better Quality of Experience (QoE) is investigated.

Fast Uplink Grant (FUG) is one of the predictive access schemes presented in 3GPP Release 14 to provide predetermined uplink allocations for IoT devices [4, 5, 52]. In [50], IoT packet transmissions have been modelled with a binary Markov process, while in [11] a FUG allocation technique was developed by combining Support Vector Machines (SVM) and Long-Short Term Memory (LSTM) neural networks.

Another trend [39, 46, 47] on predictive access schemes addresses MAP by using Joint Forecasting Scheduling (JFS), and in [39] JFS was proposed to schedule transmissions based on forecasts of generation times and sizes of bursts. JFS can be recursively enhanced with a Multi-Scale Algorithm (MSA), where the performance of JFS and the length of scheduling horizon, are significantly increased [46] (over \(96\%\) throughput) for 6400 devices with variable latency constraints. However, the computational requirements of MSA are very high and in [47] a scheduling heuristic is used to determine JFS transmission times for using multiple frequency channels.

Most recently the Randomization of Generation Times (RGT) preprocessing algorithm [36] is shown to significantly improve the performance of scheduling heuristics with very low computational cost for large numbers of IoT devices. Also, the Quasi-Deterministic Transmission Policy (QDTP) traffic shaping approach [19, 25] has been shown, using queueing theory, diffusion approximations [16] and trace driven simulations, to mitigate the MAP by drastically reducing the waiting time at gateways. This research has shown that RGT and QDTP can mitigate MAP with very low computational requirements for up to 6400 IoT devices. In the remainder of this article, we shall outline how these two avenues of recent research can alleviate MAP.

3 Mitigating MAP Using Queueing Theory and Diffusion Approximations

We will summarize together the results of two recent articles [19, 36], which offer solutions to reduce MAP. To this end, we first present the analysis in [19] of the probability that the deadline of an IoT packet is met, providing a basis for access policies in IoT networks. Then we review the RGT [36] and QDTP [19] algorithms and their performance.

In [19], the collection of IoT devices that generate packets, the communication channel, and the receiving gateway are represented as two cascaded queues [9, 32, 41, 53]:

  • The first queue translates the generation instant at the IoT device for the j-th packet \(r_j\), into its transmission instant \(t_j\), and

  • The second queue starts with the transmission instants \(t_j\) that feeds directly into the IoT gateway where the j-th packet is served in FIFO (First-In-First-Out) order with a service duration \(p_j\).

  • Note that in this case, the transmission delay within the communication channel is taken to be zero, i.e. it is assumed to be small enough to be as compared to \(p_j\) and to the durations between the other successive instants, so that the packet leaving the IoT device at time \(t_j\) arrives at the gateway at the same time instant.

Let the IoT device generats traffic in bursts of bits that are sent at the same time instant, where burst j is generated at discrete time slot \(r_j\) and should be recieved by \(d_j\). That is, we assume that there is a deadline \(\varDelta _j\) for each burst j beyond which j is of “no value”. Thus, the burst j (which can also be considered as a packet) must arrive at the receiver gateway by \(d_j=r_j + \varDelta _j\). Furthermore, the packets of various IoT devices are processed in time ordered fashion in First-In-First-Out (FIFO) order, and \(p_j\) be the “service time” during which the receiving gateway is occupied by packet j.

3.1 The Probability of Meeting Deadlines

Let the j-th packet sent from any of the IoT devices, enumerated in time order (i.e. the \(j-th\) packet is generated before the \(j+1\)-th packet), be transmitted from its IoT device exactly at the instant \(r_j\) when it is generated. Then the time spent between the generation time \(r_j\) and the time when it starts being processed at the gateway, i.e. its total waiting time, is denoted by \(W_j\), and is given by Lindley’s recursive equation [32, 53]:

$$\begin{aligned} W_{j+1} = [W_j+p_{j}-r_{j+1}+r_{j}]^+~,~j=0,1,2,~. \end{aligned}$$
(1)

where \(r_0=0\). Note that the conventional notation \([X]^+\), for a real number X, means that \([X]^+=0\) when \(X<0\), and \([X]^+=X\) if \(X\ge 0\).

Assuming that the generation times coincide with the transmission times, if the sequence generation and service times, and deadlines \(\{r_j,~p_j,~\varDelta _j\}_{j\ge 0}\) constitute a stationary random process, the probability \(\varPi _j\) that the packet j does not meet its deadline is given by:

$$\begin{aligned} \varPi _j = Prob[R_j=W_j+p_j>r_j+\varDelta _j],~and~\varPi =\lim _{j\rightarrow \infty }\varPi _j, \end{aligned}$$
(2)

where \(R_j\) is known as the response time.

Since the focus of the work in [19, 36] is on selecting the transmission instant of each packet j, denoted by \(t_j\), to minimize \(\varPi _j\) under each traffic load of the network, \(W_j\) will be replaced by a total end-to-end delay \(V_j\) to each packet, where \(D_j\) is a scheduling delay imposed to each successive packet, and \(V_j\) includes the delay at the IoT device plus the delay at the gateway:

$$\begin{aligned} V_j= & {} W^a_j+W^b_{j},~j\ge 0,~V_0=W^a_0=W^b_0=0,~and \end{aligned}$$
(3)
$$\begin{aligned} W^a_{j+1}= & {} [W^a_j+D_j-(r_{j+1}-r_j)]^+,~t_j=r_j+W^a_j, \end{aligned}$$
(4)
$$\begin{aligned} W^b_{j+1}= & {} [W^b_j+p_j-(t_{j+1}-t_j)]^+. \end{aligned}$$
(5)

In other words, we impose an initial scheduling delay \(D_j\) to each successive packet, and then consider the resulting effect on the transmission instant \(t_j\) and on the resulting delay at the IoT device \(W^a_j\) followed by the delay at the gateway \(W^b_j\). These matters are analyzed in detail in [25], wth a resulting effect on the probability of missing the deadlines:

$$\begin{aligned} \varPi ^*_j = Prob[R^*_j=V_j+p_j>r_j+\varDelta _j],~and~\varPi ^*=\lim _{j\rightarrow \infty }\varPi _j, \end{aligned}$$
(6)

when \(\{r_j,~p_j,~D_j,~\varDelta _j\}_{j\ge 0}\) constitute a stationary random process.

3.2 Interarrival and Service Time Statistics

Next, in order to compute the interarrival and service time statistics, the approach in [19] assumes that the \(p_j\)’s of all packets are independent random variables with the same distribution whose mean is E[P] and its SCV is \(C^2_B\). Moreover, \(\lambda \) denotes the interarrival rate of packets, such that \(\lambda = E[r_{j+1}-r_j])^{-1}\). It has also been assumed that the value of \(\lambda \) increases with the number of IoT devices M that are connected to the IoT gateway, and the system will operate under variable \(\lambda \) (or M) but under stable conditions, i.e. \(\lambda E[P]<1\). Also, let \(C^2_A\) denote the SCV of interarrival times of packets. Then, \(C^2_B\) and \(C^2_A\) are respectively defined as

$$\begin{aligned} C^2_B= \frac{E[P^2]}{(E[P])^2}-1,~\text {and}~C^2_A= \frac{E[(r_{j+1}-r_{j})^2]}{(E[r_{j+1}-r_j])^2}-1. \end{aligned}$$
(7)

3.3 Using the Diffusion Approximation:

Last, the diffusion approximation [13, 33] has been used in order to determine the probability that \(R_j \le \varDelta _j\), denoted by \(F_R(\varDelta )\), where \(\varDelta _j = \varDelta \) which is constant for all packets of all IoT devices. Then, the probability of missing deadline \(\varPi =1-F_R(\varDelta )\).

Subsequently, using the diffusion approximations [19] one obtains the probability density function of the response time and the probability of missing deadline as

$$\begin{aligned} f _R(t) = \int _{0}^{\infty } \frac{x}{\sqrt{2 {\pi } \alpha t^3} } e^{-\frac{(x + \beta t)^2}{2 \alpha t}}f(x)dx,~\text {then}~\varPi = 1- \int _0^{\varDelta } f_R(\tau )d\tau , \end{aligned}$$
(8)

where \(\beta =\lambda -\mu ,~~\alpha =\lambda C^2_A+\mu C^2_B\), \(\mu = 1 / E[P]\), and f(x) is yielded by the diffusion model [13].

3.4 Numerical Results Concerning the Diffusion Analysis

Now, the results concerning a publicly available IoT traffic dataset [1] present how \(\varPi \) varies with \(C^2_A\) and \(\varDelta \).

First, Fig. 1 displays \(log_{10}(\varPi )\) for \(\lambda =0.8\) and \(C^2_B=1\) and different values of \(\varDelta \) and \(C^2_A\). Note that the minimum and maximum values of \(C^2_A\) for the traffic in the dataset [1] which are 1.6 and 2.18, and the approximate value of \(C^2_A\) for the uniform distribution which will be used for the randomization policy in Sect. 3.5, are shown as vertical bars. The results in this figure show that \(\varPi \) increases with \(C^2_A\) and decreases with \(\varDelta \) when \(\lambda \) remains constant.

Then, Fig. 2 presents the values of \(\varPi \) against increasing number of devices M as well as the corresponding \(\lambda \) and \(C^2_A\). The results show that the measured value of \(\varPi \), which is the fraction of packets that do not meet their deadlines, increases as M or corresponding \(\lambda \) increases.

In summary, the results in Fig. 1 and Fig. 2 from [19] show that reducing \(C^2_A\) significantly increases the probability that any IoT traffic packet meets its deadline.

3.5 Randomization of Data Generation Times (RGT)

While the results in Sect. 2 show that fast and computationally inexpensive heuristic algorithms are promising for MAP, the work in [36] develops the RGT preprocessing algorithm which relieves the traffic load by distributing the generation times of packets over a scheduling window of duration of \(T_{sch}\).

Fig. 1.
figure 1

We show the probability of missing the deadline (y-axis) in logarithmic scale (to the base ten), estimated with the diffusion approximation. We see that it increases significantly as \(C^2_A\) increases and when the deadline \(\varDelta \) measured in slots decreases, for a fixed but high value of the arrival rate \(\lambda =0.8\). The average service rate \(\mu \) and the SCV of service time \(C^2_B\) are both fixed to 1.

In RGT, \(r_j\) is updated by adding to it an offset which is a uniformly distributed random variable (Recall that \(C^2_A = 1/3\) for uniform distribution which is shown to be significantly lower than the minimum value in the dataset) as

$$\begin{aligned} r_j^{new} \leftarrow r_j + U[\varDelta _j-S_j] \end{aligned}$$
(9)

where \(S_j\) is a safety delay that limits the upper bound of \(r_j^{new}\) such that \(0\le S_j\le \varDelta _j\) which indicates the maximum randomization and \(S_j=\varDelta _j\) indicates that there is no randomization.

Then, in [36] the value of \(S_j\) was selected by using queueing theory [17, 23] as:

$$\begin{aligned} S_j \approx min[~p_j + \frac{E[P]}{\frac{1}{\rho } - 1},~\varDelta _j~], \end{aligned}$$
(10)

where E[P] is the average processing time, and \(\rho = \lambda E[P]\).

3.6 Experimental Results Concerning RGT

We now review the performance evaluation of RGT for two known heuristic scheduling algorithms, Priority based on Average Load (PAL) [39] and non-preemptive version of Earliest Deadline First (EDF) [26], where PAL enhanced with RGT is called R-PAL and EDF enhanced with RGT is called R-EDF. The performance evaluation is performed on the publicly available dataset [1] with respect to each of throughput \(\eta \) and fraction of successfully delivered bursts \(\zeta \) metrics. Also, the performances of R-PAL and R-EDF are compared with the upper bound performances, where \(S_j=\gamma \varDelta _j\) and exhaustively search for the value of \(\gamma \) in the range [0, 1] with increments of 0.05 to maximize each of \(\eta \) and \(\zeta \).

Fig. 2.
figure 2

The probability of missing the deadline (y-axis) in logarithmic scale (to the base ten) estimated with the diffusion approximation, using the traffic statistics of the real dataset of [1], is plotted against the number of IoT devices M (x-axis) that are being used. Note that each value of M corresponds to specific measured values of \(\lambda \) and \(C^2_A\) shown along the \(x-axis\).

Accordingly, Fig. 3 displays the comparison of R-PAL and R-EDF with the upper bounds of those as well as PAL and EDF heuristics for \(\eta \) and \(\zeta \). The results in this figure show that RGT preprocessing significantly improves the throughput performance of each heuristic while the fraction of successfully delivered bursts remains the same. Furthermore, the results in Fig. 3 (top) show that the R-PAL and R-EDF significantly outperform the original versions of the heuristics PAL and EDF for a higher number of devices while both R-PAL and R-EDF are able to achieve almost the same throughput with their upper bounds. On the other hand, in Fig. 3 (bottom), one sees that \(\zeta \) is almost the same for the enhanced heuristics (R-PAL and R-EDF) and the original heuristics (PAL and EDF).

Fig. 3.
figure 3

Comparison of the upper bound R-PAL, R-PAL, PAL and the upper bound R-EDF, R-EDF, EDF algorithms for 12 to 6400 devices with respect to \(\eta \) (top) and \(\zeta \) (bottom)

3.7 The Quasi-Deterministic Transmission Policy (QDTP)

As the diffusion analysis that is discussed in Sect. 3.4 suggests, minimizing the SCV of interarrival times of traffic packets, \(C^2_A\), reduces the probability of missing deadlines of those packets. Accordingly, in [19] a “Quasi-Deterministic Transmission Policy” (QDTP) is developed to minimize the probability of missing deadlines \(\varPi \) by setting almost all of the intertransmission times to a constant D so as to reduce the value of \(C^2_A\). In QDTP whose pseudo-code is presented in Algorithm 1, \(D=\frac{1}{\lambda }\), where \(\lambda \) is the interarrival time of burst generation times.

figure a

For the practical application of QDTP, the generation times of the packets must be known in advance, similar to other predictive protocols. While advanced knowledge of the \(\lambda \) value is required, it can be easily calculated based on the creation times of the packages. Also, each IoT device must be informed of the time slot reserved for transmission of a packet created by that device before the start of the reserved slot. For this purpose, information regarding channels reserved for transmission permissions will be sent to devices via the downlink channel. Therefore, the communication channel must be bidirectional or individual IoT devices must also have the ability to detect the channel.

3.8 Experimental Results Concerning QDTP

We now examine the performance evaluation of QDTP on the same dataset [1] used for the evaluation of RGT. On the other hand, the bits are now packetized as IP packets, where the \(21-Byte\) headers (similar to the header size in LoRa-WAN) are followed by the payload, which is the bits in a burst. In addition, considering the effective data transmission rate of LoRa-WAN in free space is 5400 bit/s [54], the transmission time for the average packet length in the dataset is 33.33 ms.

Figure 4 presents the SCV of interarrival times of packets, \(C^2_A\), for both the original generation times in the dataset and the transmission time determined via QDTP. The results in this figure show that \(C^2_A\) is significantly reduced for all values of M when QDTP is used to schedule packet transmissions.

Figure 5 displays the comparison between the performance of QDTP and that of original generation times in the dataset with respect to the base 10 of the (empirically measured) probability \(\varPi \) that the deadline is missed for different values of deadline \(\varDelta \). In this figure, we see that \(\varPi \) is reduced to practically zero for \(\varDelta \ge 5\) when QDTP is used. On the other hand, when QDTP is not used (i.e. packets are transmitted when they are generated), \(\varPi \) increases with M and approaches 1.

Fig. 4.
figure 4

Measurements of the SCV of interarrival times, both for the raw IoT data from [1], and for the same data using QDTP, for a varying number of active IoT devices M. We observe that QDTP has substantially reduced the empirically measured SCV \(C^2_A\), reducing it to zero for all the distinct numbers of devices M in the dataset of [1].

Fig. 5.
figure 5

We compare the logarithm to the base 10 of the (empirically measured) probability \(\varPi \) that the deadline is missed under both the raw dataset of [1] and under the case where the QDTP is used with the same dataset for values of the deadline \(\varDelta \in \{2, 5, 10, 20\}\).

4 Conclusions

The MAP, which can occur when a massive number of devices attempt the access a single gateway, is one of the most significant challenges for future IoT networks. Due to a high latency during the transmissions of data packets caused by MAP, the deadlines for delay-critical applications can be missed. Much work has been conducted in recent years to address this problem.

This paper reviews the work on reactive and proactive approaches to MAP, showing that methods based on proactive (i.e., predictive) techniques are a highly promising avenue to mitigate MAP. The observations of the most recent research results can be recapitulated with the following remarks:

  1. 1.

    The diffusion analysis proposed in [19] for the probability of missed deadlines in MAP shows that the latency requirements of IoT devices can be met in networks with a massive number of devices by reducing the SCV of the interarrival times of packets.

  2. 2.

    The Randomization of Generation Times (RGT) preprocessing algorithm proposed by [36] significantly improves the performance of fast scheduling heuristics by randomizing generation time of packets with uniform distribution yielding an inter-arrival time SCV of close to 1/3.

  3. 3.

    The Quasi-Deterministic Transmission Policy (QDTP) [25] meets the deadlines of almost all packets for up to 6400 IoT devices by reducing the queue length at IoT gateways to nearly zero, and the SCV of packet inter-arival times is also brought down to nearly zero.