Next Article in Journal
Numerical Simulation of Swirl Flow Characteristics of CO2 Hydrate Slurry by Short Twisted Band
Next Article in Special Issue
Relationship between Age and Value of Information for a Noisy Ornstein–Uhlenbeck Process
Previous Article in Journal
ϕ-Informational Measures: Some Results and Interrelations
Previous Article in Special Issue
Scheduling Strategy Design Framework for Cyber–Physical System with Non-Negligible Propagation Delay
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Channel Quality-Based Optimal Status Update for Information Freshness in Internet of Things

School of Electronics and Information Technology, Sun Yat-sen University, Guangzhou 510006, China
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(7), 912; https://doi.org/10.3390/e23070912
Submission received: 18 June 2021 / Revised: 10 July 2021 / Accepted: 16 July 2021 / Published: 18 July 2021
(This article belongs to the Special Issue Age of Information: Concept, Metric and Tool for Network Control)

Abstract

:
This paper investigates the status updating policy for information freshness in Internet of things (IoT) systems, where the channel quality is fed back to the sensor at the beginning of each time slot. Based on the channel quality, we aim to strike a balance between the information freshness and the update cost by minimizing the weighted sum of the age of information (AoI) and the energy consumption. The optimal status updating problem is formulated as a Markov decision process (MDP), and the structure of the optimal updating policy is investigated. We prove that, given the channel quality, the optimal policy is of a threshold type with respect to the AoI. In particular, the sensor remains idle when the AoI is smaller than the threshold, while the sensor transmits the update packet when the AoI is greater than the threshold. Moreover, the threshold is proven to be a non-increasing function of channel state. A numerical-based algorithm for efficiently computing the optimal thresholds is proposed for a special case where the channel is quantized into two states. Simulation results show that our proposed policy performs better than two baseline policies.

1. Introduction

Recently, the Internet of things (IoT) has been widely used in the field of industrial manufacturing, environment monitoring, and home automation. In these applications, the sensors generate and transmit new status updates to the destination, where the freshness of the status updates is crucial for the destination to track the state of the environment and to make decisions. Thus, a new information freshness metric, namely age of information (AoI), was proposed in [1] to measure the freshness of updates from the receiver’s perspective. There are two widely used metrics, i.e., the average peak AoI [2] and the average AoI [3]. In general, the smaller the AoI is, the fresher the received updates are.
AoI was originally investigated in [1] for updating the status in vehicular networks. Considering the impact of the queueing system, the authors in [4] investigated the system performance under the M/M/1 and M/M/1/2 queueing systems with a first-come-first-served (FCFS) policy. Furthermore, the work of [5] studied how to keep the updates fresh by analyzing some general update policies, such as the zero-wait policy. The authors of [6] considered the optimal schedule problem for a more general cost that is the weighted sum of the transmission cost and the tracking inaccuracy for the information source. However, these works assumed that the communication channel is not error-prone. In practice, status updates are delivered through an erroneous wireless channel, which suffers from fading, interference, and noises. Therefore, the received updates may not be decoded correctly, which induces information aging and energy consumption.
There are several works that considered the erroneous channel [7,8]. The authors in [9] considered multiple communication channels and investigated the optimal coding and decoding schemes. The channel with an independent and identical packet error rate over time was considered in [10,11]. The work of [12] considered the impact of fading channels in packet transmission. A Markov channel was investigated in [13], where threshold policy was proven to be optimal, and a simulation-based approach was proposed to compute the corresponding threshold. However, how the information of channel quality should be exploited to improve system performance in information freshness remains to be investigated.
Channel quality indicator (CQI) feedback is commonly used in wireless communication systems [14]. In block fading channels, the channel quality, generally reported by the terminal, is highly relevant to the packet error rate (PER) [15] or, namely, the block error rate (BLER). It is probable that a received packet fails to be decoded when the channel suffers from a poor condition. However, a transmitter with the channel quality information is able to keep idle when there is deep fading, thereby saving energy. The channel quantization was also considered in [12,13], where the channel was quantized into multiple states. However, the decision making was not dependent on the channel state in [12], while [13] did not consider the freshness of information. These motivate us to introduce the information of channel quality into the design of the updating policy.
In this paper, a status update system with channel quality feedback is considered. In particular, the channel condition is quantized into multiple states, and the destination feeds the channel quality back to the sensor before the sensor updates the status. Our problem is to investigate the channel quality-based optimal status update policy, which minimizes the weighted sum of the AoI and the energy consumption. Our key contributions are summarized as follows:
  • An average cost Markov decision process (MDP) is formulated to model this problem. Due to the infinite countable states and unbounded cost of the MDP, which makes analysis difficult, the discounted version of the original problem is first investigated, and the existence of the stationary and deterministic policy to the original problem is then proven. Furthermore, it is proven that the optimal policy is a threshold structure policy with respect to the AoI for each channel state by showing the monotonic property of the value function. We also prove that the threshold is a non-increasing function of channel state.
  • By utilizing the threshold structure, a structure-aware policy iteration algorithm is proposed to efficiently obtain the optimal updating policy. Nevertheless, a numerical-based algorithm which directly computes the thresholds by non-linear fractional programming is also derived. Simulation results reveal the effects of system parameters and show that our proposed policy performs better than the zero-wait policy and periodic policy.
The rest of this paper is organized as follows. In Section 2, the system model is presented and the optimal updating problem is formulated. In Section 3, the optimal updating policy is proven to be of a threshold structure, and a threshold-based policy iteration algorithm is proposed to find the optimal policy. Section 4 presents the simulation results. Finally, we summarize our conclusions in Section 5.

2. System Model and Problem Formulation

2.1. System Description

In this paper, we consider a status update system that consists of a sensor and a destination, as shown in Figure 1. Time is divided into slots. Without loss of generality, we assume that each time slot has an equal length, which is normalized to unity. At the beginning of each slot, the destination feeds the CQI back to the sensor. It is worth noting that the PER is different for different CQIs. Based on the CQI, the sensor decides in each time slot whether it should generate and transmit a new update to the destination via a wireless channel or keep idle for saving energy. These updates are crucial for the destination to estimate the states of the surrounding environment of the sensor and to make in-time decisions. Let a t , which takes value from the action set A = { 0 , 1 } , denote the action that the sensor performs in slot t, where a t = 1 means that the sensor generates and transmits a new update to the destination, and a t = 0 represents that the sensor is idle. If the sensor transmits an update packet in slot t, an acknowledgment will be fed back at the end of this time slot. In particular, an ACK is fed back when the destination successfully receives the update packet, and a NACK otherwise.

2.2. Channel Model

Suppose that the wireless channel is a block fading channel where the channel gain remains constant in each slot and varies independently over different slots. Let z t denote the channel gain in slot t which takes value from [ 0 , + ) . We quantize the channel gain into N + 1 levels which are denoted as ( z 0 , z 1 , , z i , , z N ) . The quantization levels are arranged in an increasing order where z 0 = 0 and z N = . Hence, the channel is said to be in state i if the channel gain z t belongs to the interval [ z i , z i + 1 ) . We denote by h t the state of the channel in slot t, where h t H { 0 , 1 , 2 , N 1 } . With the aid of CQI fed back from the destination, the sensor has knowledge of the channel state at the beginning of each time slot.
Let p z ( z ) denote the distribution of the channel gain. Then, the probability of the channel being in state i is
p i = z i z i + 1 p z ( z ) d z .
We assume that the signal-to-noise ratio (SNR) per information bit during the transmission remains constant. Then, the PER depends only on the channel gain. In particular, the PER for channel state i is given by
g i = z i z i + 1 P PER ( z ) p z ( z | i ) d z ,
where P PER ( z ) is the PER of a packet with respect to the channel gain. The success probability q i of a packet transmitted over channel state i is q i = 1 g i . According to [15], the success probability is a non-decreasing function of the channel state.

2.3. Age of Information

This paper uses the AoI as the freshness metric, which is defined as the time elapsed since the generation time of the latest update packet that is successfully received by the destination [1]. Let G i be the generation time of the ith successfully received update packet. Then, the AoI in time slot t, Δ t , is defined as
Δ t = t max { G i : G i t } .
In particular, if an update packet is successfully received, the AoI decreases to one. Otherwise, the AoI increases by one. Altogether, the evolution of the AoI is expressed by
Δ t + 1 = 1 , if the transmission is successful , Δ t + 1 , otherwise .
An example of the AoI evolution is shown in Figure 2, where the gray rectangle represents a successful reception of an update packet, and the mesh rectangle represents a transmission failure.

2.4. Problem Formulation

The objective of this paper is to find an optimal updating policy that minimizes the long-term average of the weighted sum of AoI and energy consumption. A policy π can be represented by the sequence of actions, i.e., π = ( a 0 , a 1 , , a t , ) . Let Π be a set of stationary and deterministic policies. Then, the optimal updating problem is given by
min π Π lim sup T 1 T t = 0 T E [ Δ t + ω a t C e ] ,
where C e is the energy consumption, and ω is the weighting factor.

3. Optimal Updating Policy

This section aims to investigate the optimal updating policy for the problem formulated in above section. In this section, our investigating problem is first formulated into an infinite horizon average cost MDP, and the existence of a stationary and deterministic policy that minimizes the average cost is proven. Then, the non-decreasing property of the value function is derived. Based on this property, we prove that the optimal update policy is of a threshold structure with respect to AoI, and the optimal threshold is a non-increasing function of the channel state. Aiming to reduce the computational complexity, a structure-aware policy iteration algorithm is proposed to find the optimal policy. Moreover, non-linear fractional programming is employed to directly compute the optimal thresholds in a special case where the channel is quantized into two states.

3.1. MDP Formulation

The Markov decision process (MDP) is typically applied to address the optimal decision problem when the investigation problem can be characterized by the evolution of the system state and the cost is per-stage. The optimization problem in (5) can be formulated as an infinite horizon average cost MDP, which is elaborated in the following.
  • States: The state of the MDP in slot t is defined as x t = ( Δ t , h t ) , which takes values in Z + × H . Hence, the state space S is countable and infinite.
  • Actions: The set of actions a t chosen in slot t is A = { 0 , 1 } .
  • Transition Probability: Let Pr ( x t + 1 | x t , a t ) be the transition probability that the state x t in slot t transits to x t + 1 in slot t + 1 after taking action a t . According to the evolution of AoI in (4), the transition probability is given by
    Pr ( x t + 1 = ( Δ + 1 , j ) | x t = ( Δ , i ) , a t = 0 ) = p j , Pr ( x t + 1 = ( 1 , j ) | x t = ( Δ , i ) , a t = 1 ) = q i p j , Pr ( x t + 1 = ( Δ + 1 , j ) | x t = ( Δ , i ) , a t = 1 ) = g i p j .
  • Cost: The instantaneous cost C ( x t , a t ) at state x t given action a t in slot t is
    C ( x t , a t ) = Δ t + ω a t C e .
For an MDP with infinite states and unbounded cost, it is not guaranteed to have a stationary and deterministic policy that attains the minimum average cost in general. Fortunately, we can prove the existence of stationary and deterministic policy in next sub-section.

3.2. The Existence of Stationary and Deterministic Policy

For rigorous mathematical analysis, this section is purposed to prove the existence of a stationary and deterministic optimal policy. According to [16], we first analyze the associated discounted cost problem of the original MDP. The expectation of discount cost with respect to discounted factor γ and initial state x ^ under a policy π is given by
V π , γ ( x ^ ) = E π t = 0 γ t C ( x t , a t ) | x 0 = x ^ ,
where a t is the decision made in state x ^ under policy π , and γ ( 0 , 1 ) is the discounted factor. We first verify that V π , γ ( x ^ ) is finite for any policy and all x ^ S .
Lemma 1.
Given γ ( 0 , 1 ) , for any policy π and all x ^ = ( x ^ , h ^ ) S , we have
V π , γ ( x ^ ) = E π t = 0 γ t C ( x t , a t ) | x 0 = x ^ < .
Proof. 
By definition, the instantaneous cost in state x t = ( Δ t , h t ) given action a t is
C ( x t , a t ) = Δ t , if a t = 0 , Δ t + ω C e , if a t = 1 .
Therefore, C ( x t , a t ) Δ t + ω C e holds. Combined with the fact that the AoI increases, at most, linearly at each slot for any policy, we have
t = 0 γ t C ( x t , a t | x 0 = ( Δ ^ , h ^ ) ) t = 0 γ t ( Δ ^ + t + ω C e ) = 1 1 γ Δ ^ + γ 1 γ + ω < ,
which completes the proof.    □
Let V γ ( x ^ ) = min π V π , γ ( x ^ ) denote the minimum expected discounted cost. By Lemma 1, V γ ( x ^ ) = min π V π , γ ( x ^ ) < holds for every x ^ and γ ( 0 , 1 ) .
According to [16] (Proposition 1), we have
V γ ( x ^ ) = min a A C ( x ^ , a ) + γ x S Pr ( x | x ^ , a ) V γ ( x ) ,
which implies that V γ ( x ^ ) satisfies the Bellman equation. V γ ( x ^ ) can be solved via a value iteration algorithm. In particular, we define V γ , 0 ( x ^ ) = 0 , and for all n 1 , we have
V γ , n ( x ^ ) = min a A Q γ , n ( x ^ , a ) ,
where
Q γ , n ( x ^ , a ) = min a A C ( x ^ , a ) + γ x S Pr ( x | x ^ , a ) V γ , n 1 ( x )
is related to the right-hand-side (RHS) of the discounted cost optimality equation. Then, lim n V γ , n ( x ^ ) = V γ ( x ^ ) for every x ^ and γ .
Now, we can use the value iteration algorithm to establish the monotonic properties of V γ ( x ^ )
Lemma 2.
For all Δ and i, we have
V γ ( Δ , N 1 ) V γ ( Δ , i ) ,
and for all Δ 1 Δ 2 and i, we have
V γ ( Δ 1 , i ) V γ ( Δ 2 , i ) .
Proof. 
See Appendix A.    □
Based on Lemmas 1 and 2, we are ready to show that the MDP has a stationary and deterministic optimal policy in the following theorem.
Theorem 1.
For the MDP in (5), there exists a stationary and deterministic optimal policy π * that minimizes the long-term average cost. Moreover, there exists a finite constant λ = lim γ 1 ( 1 γ ) V γ ( x ) for all states x , where λ is independent of the initial state, and a value function V ( x ) , such that
λ + V ( x ) = min a A C ( x , a ) + x S Pr ( x | x , a ) V ( x )
holds for all x .
Proof. 
See Appendix B.    □

3.3. Structural Analysis

According to Theorem 1, the optimal policy for the average cost problem satisfies the following equation
π * ( x ) = arg min a A Q ( x , a ) ,
where
Q ( x , a ) = C ( x , a ) + x S Pr ( x | x , a ) V ( x ) .
Similar to Lemma 2, the monotonic property of the value function V ( x ) is given in the following lemma.
Lemma 3.
Given the channel state i, for any Δ 2 Δ 1 , we have
V ( Δ 2 , i ) V ( Δ 1 , i ) .
Proof. 
This proof follows the same procedure of Lemma 2, with one exception being that the value iteration algorithm is based on Equation (17).    □
Moreover, based on Lemma 3, the property of the increment of the value function is established in following lemma.
Lemma 4.
Given the channel state i, for any Δ 2 Δ 1 , we have
V ( Δ 2 , i ) V ( Δ 1 , i ) Δ 2 Δ 1 .
Proof. 
We first examine the relation between the state-action value functions, i.e., Q ( Δ 2 , i , a ) and Q ( Δ 1 , i , a ) . Specifically, based on Lemma 3, we have
Q ( Δ 2 , i , 0 ) ( Δ 2 Δ 1 ) = Δ 1 + j = 0 N 1 p j V ( Δ 2 + 1 , j ) Δ 1 + j = 0 N 1 p j V ( Δ 1 + 1 , j ) = Q ( Δ 1 , i , 0 ) ,
and
Q ( Δ 2 , i , 1 ) ( Δ 2 Δ 1 ) = Δ 1 + ω C e + q i j = 0 N 1 p j V ( 1 , j ) + g i j = 0 N 1 p j V ( Δ 2 + 1 , j ) Δ 1 + ω C e + q i j = 0 N 1 p j V ( 1 , j ) + g i j = 0 N 1 p j V ( Δ 1 + 1 , j ) = Q ( Δ 1 , i , 1 ) .
Since V ( x ) = min a A Q ( x , a ) , we complete the proof.    □
Our main result is presented in the following theorem.
Theorem 2.
For any given channel state i, there exists a threshold β i , such that when Δ β i , the optimal action is to generate and transmit a new update, i.e., π * ( Δ , i ) = 1 , and when Δ < β i , the optimal action is to remain idle, i.e., π * ( Δ , i ) = 0 . Moreover, the optimal threshold β i is a non-increasing function of channel state i, i.e., β i β j holds for all i , j H and i j .
Proof. 
See Appendix C.    □
According to Theorem 2, the sensor will not update the status until the AoI exceeds the threshold. Moreover, if the channel condition is not good, i.e., channel state i is small, the sensor will wait for a longer time before it samples and transmits the status update packet so as to reduce the energy consumption because of a higher probability of transmission failure.
Based on the threshold structure, we can reduce the computational complexity of the policy iteration algorithm to find the optimal policy. The details of the algorithm are presented in Algorithm 1.
Algorithm 1 Policy iteration algorithm (PIA) based on the threshold structure.
1:
Initialization: Set k = 0 , iteration threshold ϵ , and initialize the value function V 0 ( x ) = 0 and policy π ( x ) = 0 for all state x S
2:
repeat
3:
k k + 1 .
4:
  Based on last iterative value function V k 1 ( x ) , compute the current value function V k ( x ) by calculating the following equations.
5:
V k ( x ) = min a A C ( x , a ) + x S Pr ( x | x , a ) V k 1 ( x )
6:
until | V k ( x ) V k 1 ( x ) | ϵ for all x S
7:
for x = ( Δ , i ) S do
8:
if x = ( Δ 1 , i ) S and π ( x ) = 1  then
9:
   π ( x ) 1 .
10:
else
11:
    π ( x ) arg min a A C ( x , a ) + Pr x S ( x | x , a ) V ( x )
12:
end if
13:
end for
14:
π * π
15:
return the optimal policy π *

3.4. Computing the Thresholds for a Special Case

In the above section, we have proven that the optimal policy has a threshold structure. Given the thresholds ( β 0 , β 1 , , β N 1 ) , a Markov chain can be induced by the threshold policy. A special Markov chain is depicted in Figure 3, where the channel has two states. By leveraging the Markov chain, we first derive the average cost of the special case, which is summarized in the following theorem.
Theorem 3.
Let φ ( x ) be the steady state probability of state x of the corresponding Markov chain with two states and β 0 , β 1 be the threshold with respect to the channel state, respectively. The steady state probability is given by
φ ( i , j ) = p j φ 1 , if 1 i β 1 , p j s 0 i β 1 φ 1 , if β 1 < i β 0 , p j s 0 β 0 β 1 s 1 i β 1 φ 1 , if i > β 0 ,
where φ 1 = φ ( 1 , 0 ) + φ ( 1 , 1 ) , s 0 = 1 p 1 q 1 , s 1 = 1 p 0 q 0 p 1 q 1 , and φ 1 satisfies following equation:
φ 1 = 1 β 1 + s 0 s 0 β 0 β 1 1 s 0 + s 0 β 0 β 1 s 1 1 s 1 .
The average cost then is given by
C m c ( β 0 , , β N 1 ) = φ 1 ( β 1 ( β 1 + 1 ) 2 + A + B + ω C e E ) ,
where
A = s 0 ( ( β 1 + 1 ) β 0 s 0 β 0 β 1 ) 1 s 0 + s 0 3 s 0 β 0 β 1 + 1 ( 1 s 0 ) 2 ,
B = ( β 0 + 1 ) s 1 1 s 1 + s 1 3 1 s 1 ,
and
E = s 0 β 0 β 1 1 1 s 1 + p 1 1 s 0 β 0 β 1 1 s 0 .
Proof. 
See Appendix D.    □
Therefore, the closed form of the average cost is a function of thresholds. By linear search or gradient descent algorithm, the numerical solution of optimal thresholds can be obtained. However, computing its gradient directly requires a large amount of computation till convergence. Here, a nonlinear fractional programming (NLP) [17] based algorithm which can efficiently obtain the numerical solution is proposed.
Let x = ( β 0 , β 1 ) . We can rewrite the cost function as a fractional form, where the numerator is denoted as N ( x ) = C m c ( x ) / φ 1 , and the denominator term is N ( x ) = 1 / φ 1 . The solution to an NLP problem with the form in the following
max N ( x ) D ( x ) | x A
is related to the optimization problem (31)
max N ( x ) q D ( x ) | ( x A , for q R ,
where the following assumption should also be satisfied:
D ( x ) > 0 , for all x A .
Define the function F ( q ) with variable q as
F ( q ) = max N ( x ) q D ( x ) | x A , for q R .
According to [17], F ( q ) is a strictly monotonic decreasing function and is convex over R . Furthermore, we have q 0 = N ( x 0 ) / D ( x 0 ) = max { N ( x ) q D ( x ) | x A } if, and only if,
F ( q 0 ) = max { N ( x ) q 0 D ( x ) | x A } = 0 .
Then, the algorithm can be described by two steps. The first step is to solve a convex optimization problem with a one dimensional parameter by a bisection method. The second step is to solve a high dimensional optimization problem by a gradient descent method.
According to [17], a bisection method can be used to solve the optimal q 0 , under the assumption that the value of function F ( q ) can be obtained exactly for given q. We will actually use the gradient descent algorithm to obtain the numerical solution of F ( q ) since the global search method may not perform in polynomial time. As a trick, we alternate the optimization variables of thresholds ( β 0 , β 1 ) by the variables of the decrement of thresholds, i.e., x = ( β 0 β 1 , β 1 ) . To summarize, the numerical-based method for computing the optimal thresholds is given by Algorithm 2.
Algorithm 2 Numerical computation of the optimal thresholds.
Input: Iteration time k , error threshold δ
Output: Numerical result x *
1:
Let N ( x ) = C m c ( x ) / φ 1 , and D ( x ) = 1 / φ 1 . Define F ( q ) = max { N ( x ) q D ( x ) | x 0 }
2:
Let the iteration starts with i = 1 , search range [ a , b ] of q.
3:
while i k do
4:
m = a + b 2 ;
5:
  if F ( m ) F ( a ) < 0 then
6:
b = m ;
7:
  else
8:
a = m ;
9:
  end if
10:
 if b a 2 < δ then
11:
     x * = arg min x F ( m )
12:
   break;
13:
 end if
14:
    i = i + 1 ;
15:
end while

4. Simulation Results and Discussions

In this section, the simulation results are presented to investigate the impacts of the system parameters. We also compare the optimal policy with the zero-wait policy and periodic policy, where the zero-wait policy immediately generates an update at each time slot and the periodic policy keeps a constant interval between two updates.
Figure 4 depicts the optimal policy for different AoI and channel states, where the number of channel states is 5. It can be seen that, for each channel state, the optimal policy has a threshold structure with respect to the AoI. In particular, when the AoI is small, it is not beneficial for the sensor to generate and transmit a new update because the energy consumption dominates the total cost. We can also see that the threshold is non-increasing with the channel state. In other words, if the channel condition is better, the threshold is smaller. This is because the success probability of packet transmission increases with the channel state.
Figure 5 illustrates the thresholds for the MDP with two channel states with respect to the weighting factor ω , in which the two dashed lines are obtained by PIA and the other two solid lines are obtained by the proposed numerical algorithm. Both of the thresholds grow with the increasing of ω . Since the energy consumption has more weight, it is not efficient to update when the AoI is small. On the contrary, when ω decreases, the AoI dominates and the thresholds decline. In particular, both of the thresholds equal 1 when ω = 0 . In this case, the optimal policy reduces to the zero-wait policy. We can also see that the value of the threshold for channel state 1 of the numerical algorithm is close to the optimal solution. In contrast, the value of the threshold for channel state 0 gradually deviates from the optimal value.
Figure 6 illustrates the performance comparison of four policies, i.e., the zero-wait policy, the periodic policy, the numerical-based policy, and the optimal policy, with respect to the weighting factor ω . It is easy to see that the optimal policy has the lowest average cost. As we see in Figure 6, the zero-wait policy has the same performance with the optimal policy when ω = 0 . As ω increases, the average cost of all three policies increases. However, the increment of the zero-wait policy is larger than the periodic policy and the optimal policy due to the frequent transmission in the zero-wait policy. Although the thresholds obtained by the PIA and the numerical algorithm are not exactly the same as shown in Figure 5, the performance of the numerical-based algorithm also coincides with the optimal policy. This is because the threshold for channel state 1 exists in the quadratic term of the cost function, while the threshold for channel state 0 exists in the negative exponential term of the cost function. As a result, the threshold for channel state 1 has a much more significant effect on the system performance.
Figure 7 compares the three policies with respect to the probability p 1 of the channel being in state 1. Since there is a higher probability that the channel has a good quality as p 1 increases, the average cost of all three policies decreases. We can see that, in the regime of p 1 , the optimal policy has the lowest average cost, because it can achieve a good balance between the AoI and the energy consumption. We can also see that the cost of the periodic policy is greater than the zero-wait policy first, and smaller later. To further demonstrate these curves, we separate the energy consumption term and AoI term into different figures, i.e., Figure 8 and Figure 9. We see that the update cost of the zero-wait policy is smaller than that of the periodic policy, but the AoI of the zero-wait policy has a smaller decrease with respect to p 1 than the periodic policy.

5. Conclusions

In this paper, we have studied the optimal updating policy in an IoT system, where the channel gain is quantized into multiple states and the channel state is fed back to the sensor before the decision making. The status update problem has been formulated as an MDP to minimize the long-term average of the weighted sum of the AoI and the energy consumption. By investigating the properties of the value function, it is proven that the optimal policy has a threshold structure with respect to AoI for any given channel state. We have also proven that the threshold is a non-increasing function of the channel state. Simulation results show the impacts of system parameters on the optimal thresholds and the average cost. Through comparisons, we have also shown that our proposed policy outperforms the zero-wait policy and the periodic policy. In our future research, the time-varying channel model will be further involved for guiding the future design of realistic IoT systems.

Author Contributions

Conceptualization, F.P., X.C., and X.W.; methodology, F.P. and X.W.; software, F.P.; validation, F.P., X.C., and X.W.; formal analysis, F.P. and X.W.; investigation, X.W.; writing—original draft preparation, F.P.; writing—review and editing, X.W. and X.C.; visualization, F.P. and X.W.; supervision, X.C. and X.W.; project administration, X.C.; funding acquisition, X.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the State’s Key Project of Research and Development Plan under Grants (2019YFE0196400), in part by Guangdong R&D Project in Key Areas under Grant (2019B010158001), in part by Guangdong Basic and Applied Basic Research Foundation under Grant (2021A1515012631), in part by Key Laboratory of Modern Measurement & Control Technology, Ministry of Education, Beijing Information Science & Technology University (KF20201123202).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Proof of Lemma 2

Based on the value iteration algorithm, the induction method can be employed in following proof. Firstly, we initial that V γ , 0 ( x ) = 0 , where both Equations (15) and (16) hold for all x S .

Appendix A.1. Proof of Equation (16)

When n = 1 ,
Q γ , 1 ( Δ 1 , i , 0 ) Q γ , 1 ( Δ 2 , i , 0 ) = Δ 1 Δ 2 0 ,
and
Q γ , 1 ( Δ 1 , i , 1 ) Q γ , 1 ( Δ 2 , i , 1 ) = Δ 1 + ω C e ( Δ 2 + ω C e ) 0 ,
hold due to Δ 1 Δ 2 , and we have V γ , 1 ( Δ 1 , i ) V γ , 1 ( Δ 2 , i ) .
Suppose that V γ , K ( Δ 1 , i ) V γ , K ( Δ 2 , i ) holds for k K . Considering the case of k = K + 1 ,
Q γ , K + 1 ( Δ 1 , i , 0 ) Q γ , K + 1 ( Δ 2 , i , 0 ) = Δ 1 Δ 2 0 ,
and
Q γ , K + 1 ( Δ 1 , i , 1 ) Q γ , K + 1 ( Δ 2 , i , 1 ) = ( Δ 1 Δ 2 ) + γ j H p j g i V γ , K ( Δ 1 + 1 , j ) V γ , K ( Δ 2 + 1 , j ) 0 ,
hold for all i according to Δ 1 Δ 2 . Therefore, we have V γ , K + 1 ( Δ 1 , i ) V γ , K + 1 ( Δ 2 , i ) . Since lim n V γ , n = V γ , we have V γ ( Δ 1 , i ) V γ ( Δ 2 , i ) .

Appendix A.2. Proof of Equation (15)

By the definition of function Q γ ( x , a ) , we have
Q γ ( Δ , i , 0 ) = Δ + γ j H p j V γ ( Δ + 1 , j ) ,
and
Q γ ( Δ , i , 1 ) = Δ + ω C e + γ j H p j g i V γ ( Δ + 1 , j ) + q i V γ ( 1 , j ) .
Therefore,
Q γ ( Δ , N 1 , 0 ) Q γ ( Δ , i , 0 ) 0 ,
and
Q γ ( Δ , N 1 , 1 ) Q γ ( Δ , i , 1 ) = ( a ) γ j H p j ( q N 1 q i ) V γ ( 1 , j ) V γ ( Δ + 1 , j ) 0 ,
hold for all i, where step ( a ) is due to Equation (16). Hence, we have V γ ( Δ , N 1 ) V γ ( Δ , i ) . This completes the whole proof.

Appendix B. Proof of Theorem 1

Theorem 1 can be proven by verifying the conditions given in [16]. The conditions are listed as follows:
  • (1): For every state x and discount factor γ , the discount value function V γ ( x ) is finite.
  • (2): There exists a non-negative value L such that L h γ ( x ) for all x and γ , where h γ ( x ) = V γ ( x ) V γ ( x ^ ) , and x ^ is a reference state.
  • (3): There exists a non-negative value M x , such that h γ ( x ) M x for every x and γ . For every x , there exists an action a x such that x Pr ( x | x , a x ) M x < .
  • (4): The inequality x Pr ( x | x , a ) M x < holds for all x and a.
By Lemma 1, V γ ( x ^ ) = min π V π , γ ( x ^ ) < holds for every x ^ and γ . Hence, condition (1) holds. According to Lemma 2, by letting x ^ = ( 1 , N 1 ) and L = 0 , we have h γ ( x ) 0 , which verifies condition (2).
Before verifying condition (3), a lemma is given as follows:
Lemma A1.
Let us denote x ^ = ( 1 , N 1 ) as the reference state and define the first time that an initial state x transits to x ^ as K = min k : k 1 , x k = x ^ . Then, the expectation cost under the always-transmitting policy π a , i.e., the sensor generates and transmits a new update in each slot, is
C x , x ^ ( π a ) = E π a t = 0 K 1 γ t C ( x t , a t ) | x ,
where C x , x ^ ( π a ) < holds for all x .
Proof. 
Since a t = 1 for all t, the probability that the state returns to x ^ from x after exactly K slot is given by
Pr ( K = k | x = ( Δ , j ) ) = p N 1 q j , if k = 1 , p N 1 g j ( 1 i H p i q i ) k 2 ( i H p i q i ) , otherwise .
Then, the expectation return cost from x to x ^ is expressed as
C x , x ^ ( π a ) = E t = 0 K 1 γ t C ( x t , a t ) | x ( a ) k = 1 Pr K = k | x = ( Δ , j ) m = 0 k 1 ( Δ + m + ω C e ) < ,
where step ( a ) is due to the fact that C ( x t , a t ) Δ t + ω C e . □
Considering a mixture policy π , in which it performs the always-transmitting policy π a from initial state x until it enters the reference state x ^ , it later performs the optimal policy π γ that minimizes the discounted cost. Therefore, we have
V γ ( x ) E π a t = 0 K 1 γ t C ( x t , a t ) | x + E π b t = K γ t C ( x t , a t ) | x ^ C x , x ^ ( π a ) + E π γ K V γ ( x ^ ) C x , x ^ ( π a ) + V γ ( x ^ ) ,
which implies that h γ ( x ) C x , x ^ ( π a ) . Hence, let x ^ = ( 1 , N 1 ) and M x = C x , x ^ ( π a ) ; condition (3) is verified.
On the other hand, M x < holds for all x . The states that transit from x are finite. Thus, the weighted sum of finite M x is also finite, i.e., x P ( x | x , a ) M x < holds for all x and a, which verifies condition (4). This completes the whole verification.

Appendix C. Proof of Theorem 2

Based on the definition of Q ( Δ , i , a ) , we can obtain the difference between the state-action value function as follows:
Q ( Δ , i , 0 ) Q ( Δ , i , 1 ) = j = 0 N 1 p j V ( Δ + 1 , j ) q i j = 0 N 1 p j V ( 1 , j ) g i j = 0 N 1 p j V ( Δ + 1 , j ) ω C e = q i j = 0 N 1 p j V ( Δ + 1 , j ) V ( 1 , j ) ω C e ( a ) q i Δ ω C e .
where ( a ) is due to the property of the value function given in Lemma 4. We then discuss the difference between the state-action value function in two cases.
Case 1: ω = 0 .
In this case, Q ( Δ , i , 0 ) Q ( Δ , i , 1 ) 0 holds for any Δ and i. Therefore, the optimal policy is to update at each slot in spite of the channel state. In other words, the optimal thresholds are all equal to 1.
Case 2: ω > 0 .
We note that, given i, q i Δ ω C e increases linearly with Δ . Hence, there exists a positive integer β ^ i , such that β ^ i is the minimum value that satisfies q i β ^ i ω C e 0 . Therefore, if Δ β ^ i , Q ( Δ , i , 0 ) Q ( Δ , i , 1 ) q i β ^ i ω C e 0 holds. This implies that there must exist a threshold β i satisfying 1 β i β ^ i . If Δ β i , we have Q ( Δ , i , 0 ) Q ( Δ , i , 1 ) 0 .
Altogether, the optimal policy has a threshold structure for ω 0 . Then, we examine the non-increasing property of the thresholds. Firstly, we show that the difference between the state-action value function is monotonic with respect to the channel state by fixing the AoI. Assuming that i , j H and i j , it is easy to obtain that
Q ( Δ , j , 0 ) Q ( Δ , j , 1 ) ( Q ( Δ , i , 0 ) Q ( Δ , i , 1 ) ) = ( q j q i ) l = 0 N 1 p l ( V ( Δ + 1 , l ) V ( 1 , l ) ) 0 .
Since Q ( Δ , i , 0 ) Q ( Δ , i , 1 ) 0 when Δ β i , we have Q ( Δ , j , 0 ) Q ( Δ , j , 1 ) 0 according to (A14). This implies that the optimal threshold β j corresponding to channel state j is no greater than β i , i.e., β j β i . This completes the whole proof.

Appendix D. Proof of Theorem 3

Assume that φ ( x ) is the steady probability of state x in a Markov chain. The steady state probability φ ( x ) satisfies the following global balance equation [18], i.e.,
φ ( x ) = x S φ ( x ) Pr ( x | x ) .
Let φ 1 = φ ( 1 , 0 ) + φ ( 1 , 1 ) . We prove Equation (A23) by discussing three cases via mathematical induction.
Case 1: 1 < i β 1
Based on Equation (A15), we have
φ ( 2 , j ) = φ ( 1 , 0 ) p j + φ ( 1 , 1 ) p j = p j φ 1 .
Assuming that φ ( i , j ) = p j φ 1 holds for all i k < β 1 , we examine φ ( k + 1 , j ) . We have
φ ( k + 1 , j ) = φ ( k , 0 ) p j + φ ( k , 1 ) p j = p j p 0 φ 1 + p j p 1 φ 1 = p j φ 1 ,
which completes this segment of the proof.
Case 2: β 1 < i β 0
Similarly, we have
φ ( β 1 + 1 , j ) = p j ( 1 q 1 ) φ ( β 1 , 1 ) + p j φ ( β 1 , 0 ) = p j φ 1 ( 1 q 1 ) p 1 + p 0 = p j φ 1 s 0 ,
where s 0 = 1 p 1 q 1 . Assuming that φ ( i , j ) = p j φ 1 s 0 i β 1 holds for all β 1 < i k < β 0 , we examine φ ( k + 1 , j ) . We have
φ ( k + 1 , j ) = p j ( 1 q 1 ) φ ( k , 1 ) + p j φ ( k , 0 ) = p j φ 1 ( 1 q 1 ) p 1 + p 0 s 0 k β 1 = p j φ 1 s 0 k + 1 β 1 ,
which completes this segment of the proof.
Case 3: i > β 0
Follow above discussion, we have
φ ( β 0 + 1 , j ) = p j ( 1 q 1 ) φ ( β 0 , 1 ) + p j ( 1 q 0 ) φ ( β 0 , 0 ) = p j φ 1 s 0 β 0 β 1 ( 1 q 1 ) p 1 + ( 1 q 0 ) p 0 = p j φ 1 s 0 β 0 β 1 s 1 ,
where s 1 = 1 p 0 q 0 p 1 q 1 . Assuming that φ ( i , j ) = p j φ 1 s 0 β 0 β 1 s 1 i β 0 holds for all β 0 < i k , we examine φ ( k + 1 , j ) . We have
φ ( k + 1 , j ) = p j ( 1 q 1 ) φ ( k , 1 ) + p j ( 1 q 0 ) φ ( k , 0 ) = p j φ 1 s 0 β 0 β 1 ( 1 q 1 ) p 1 + ( 1 q 0 ) p 0 s 1 k β 0 = p j φ 1 s 0 β 0 β 1 s 1 k + 1 β 0 .
Altogether, we obtain the steady state probability with respect to an unknown parameter φ 1 . According to the fact that i = 1 j = 0 N 1 φ ( i , j ) = 1 , we formulate an equation:
i = 1 j = 0 N 1 φ ( i , j ) = φ 1 β 1 + i = β 1 + 1 β 0 s 0 i β 1 + i = β 0 + 1 s 0 β 0 β 1 s 1 i β 0 = φ 1 β 1 + s 0 s 0 β 1 β 0 1 s 0 + s 0 β 1 β 0 s 1 1 s 1 = 1 ,
where the expression of φ 1 is obtained.
The average cost of a Markov chain is given by
C m c = x S φ ( x ) C ( x , π * ( x ) ) .
Substituting (24) into (A23), we have
C m c ( β 0 , β 1 ) = i = 1 i j = 0 1 φ ( i , j ) + i = β 0 ω C e φ ( i , 0 ) + i = β 1 ω C e φ ( i , 1 ) .
Furthermore, the first term is given by
i = 1 i j = 0 1 φ ( i , j ) = i = 1 β 1 i φ i + i = β 1 + 1 β 0 i s 0 i β 1 + s 0 β 0 β 1 i = β 0 + 1 i s 1 i β 0 = φ 1 β 1 ( β 1 + 1 ) 2 + φ 1 A + φ 1 B ,
where
A = s 0 ( ( β 1 + 1 ) β 0 s 0 β 0 β 1 ) 1 s 0 + s 0 3 s 0 β 0 β 1 + 1 ( 1 s 0 ) 2 ,
and
B = ( β 0 + 1 ) s 1 1 s 1 + s 1 3 1 s 1 .
Furthermore, the sum of last two terms is given by
i = β 0 ω C e φ ( i , 0 ) + i = β 1 ω C e φ ( i , 1 ) = φ 1 ω C e s 0 β 0 β 1 i = β 0 s 1 i β 0 + p 1 i = β 1 β 0 1 s 0 i β 1 = φ 1 ω C e s 0 β 0 β 1 1 1 s 1 + p 1 1 s 0 β 0 β 1 1 s 0 .
This completes the proof.

References

  1. Kaul, S.; Yates, R.; Gruteser, M. Real-time status: How often should one update? In Proceedings of the IEEE INFOCOM, Orlando, FL, USA, 25–30 March 2012; pp. 2731–2735. [Google Scholar]
  2. Abd-Elmagid, M.A.; Dhillon, H.S. Average peak age-of-information minimization in UAV-assisted IoT networks. IEEE Trans. Veh. Technol. 2018, 68, 2003–2008. [Google Scholar] [CrossRef] [Green Version]
  3. Farazi, S.; Klein, A.G.; Brown, D.R. Average age of information for status update systems with an energy harvesting server. In Proceedings of the IEEE INFOCOM 2018-IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Honolulu, HI, USA, 15–19 April 2018; pp. 112–117. [Google Scholar]
  4. Costa, M.; Codreanu, M.; Ephremides, A. On the age of information in status update systems with packet management. IEEE Trans. Inf. Theory 2016, 62, 1897–1910. [Google Scholar] [CrossRef]
  5. Sun, Y.; Uysal-Biyikoglu, E.; Yates, R.D.; Koksal, C.E.; Shroff, N.B. Update or wait: How to keep your data fresh. IEEE Trans. Inf. Theory 2017, 63, 7492–7508. [Google Scholar] [CrossRef]
  6. Yun, J.; Joo, C.; Eryilmaz, A. Optimal real-time monitoring of an information source under communication costs. In Proceedings of the 2018 IEEE Conference on Decision and Control (CDC), Miami, FL, USA, 17–19 December 2018; pp. 4767–4772. [Google Scholar]
  7. Imer, O.C.; Basar, T. Optimal estimation with limited measurements. In Proceedings of the 44th IEEE Conference on Decision and Control, Seville, Spain, 15 December 2005; pp. 1029–1034. [Google Scholar]
  8. Rabi, M.; Moustakides, G.V.; Baras, J.S. Adaptive sampling for linear state estimation. Siam J. Control Optim. 2009, 50, 672–702. [Google Scholar] [CrossRef]
  9. Gao, X.; Akyol, E.; Başar, T. On remote estimation with multiple communication channels. In Proceedings of the 2016 American Control Conference (ACC), Boston, MA, USA, 6–8 July 2016; pp. 5425–5430. [Google Scholar]
  10. Chakravorty, J.; Mahajan, A. Remote-state estimation with packet drop. IFAC-PapersOnLine 2016, 49, 7–12. [Google Scholar] [CrossRef]
  11. Lipsa, G.M.; Martins, N.C. Optimal state estimation in the presence of communication costs and packet drops. In Proceedings of the 2009 47th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 30 September–2 October 2009; pp. 160–169. [Google Scholar]
  12. Huang, H.; Qiao, D.; Gursoy, M.C. Age-energy tradeoff in fading channels with packet-based transmissions. In Proceedings of the IEEE INFOCOM 2020-IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Toronto, ON, Canada, 6–9 July 2020; pp. 323–328. [Google Scholar]
  13. Chakravorty, J.; Mahajan, A. Remote estimation over a packet-drop channel with Markovian state. IEEE Trans. Autom. Control 2019, 65, 2016–2031. [Google Scholar] [CrossRef] [Green Version]
  14. Chen, X.; Yi, H.; Luo, H.; Yu, H.; Wang, H. A novel CQI calculation scheme in LTE∖LTE-A systems. In Proceedings of the 2011 International Conference on Wireless Communications and Signal Processing (WCSP), Nanjing, China, 9–11 November 2011; pp. 1–5. [Google Scholar]
  15. Toyserkani, A.T.; Strom, E.G.; Svensson, A. An analytical approximation to the block error rate in Nakagami-m non-selective block fading channels. IEEE Trans. Wirel. Commun. 2010, 9, 1543–1546. [Google Scholar] [CrossRef] [Green Version]
  16. Sennott, L.I. Average cost optimal stationary policies in infinite state Markov decision processes with unbounded costs. Oper. Res. 1989, 37, 626–633. [Google Scholar] [CrossRef]
  17. Dinkelbach, W. On nonlinear fractional programming. Manag. Sci. 1967, 13, 492–498. [Google Scholar] [CrossRef]
  18. Chandry, K.M. The Analysis and Solutions for General Queueing Networks; Citeseer: Princeton, NJ, USA, 1974. [Google Scholar]
Figure 1. System model.
Figure 1. System model.
Entropy 23 00912 g001
Figure 2. An example of the AoI evolution with the channel state h t , the action a t , and the acknowledgment ACK t . The asterisk stands for no acknowledgment from destination when the sensor keeps idle.
Figure 2. An example of the AoI evolution with the channel state h t , the action a t , and the acknowledgment ACK t . The asterisk stands for no acknowledgment from destination when the sensor keeps idle.
Entropy 23 00912 g002
Figure 3. An illustration of established Markov chain with two channel states.
Figure 3. An illustration of established Markov chain with two channel states.
Entropy 23 00912 g003
Figure 4. Optimal policy for different AoI and channel states ( q 0 = 0.1 ,   q 1 = 0.2 ,   q 2 = 0.3 ,   q 3 = 0.4 ,   q 4 = 0.5 ,   p 0 = 0.1 ,   p 1 = 0.1 ,   p 2 = 0.3 ,   p 3 = 0.3 ,   p 4 = 0.2 ,   ω = 10 ,   C e = 1 ).
Figure 4. Optimal policy for different AoI and channel states ( q 0 = 0.1 ,   q 1 = 0.2 ,   q 2 = 0.3 ,   q 3 = 0.4 ,   q 4 = 0.5 ,   p 0 = 0.1 ,   p 1 = 0.1 ,   p 2 = 0.3 ,   p 3 = 0.3 ,   p 4 = 0.2 ,   ω = 10 ,   C e = 1 ).
Entropy 23 00912 g004
Figure 5. Optimal thresholds for two different channel states versus ω ( p 0 = 0.2 ,   p 1 = 0.8 ,   q 0 = 0.2 ,   q 1 = 0.5 ,   C e = 1 ).
Figure 5. Optimal thresholds for two different channel states versus ω ( p 0 = 0.2 ,   p 1 = 0.8 ,   q 0 = 0.2 ,   q 1 = 0.5 ,   C e = 1 ).
Entropy 23 00912 g005
Figure 6. Comparison of the zero-wait policy, the periodic policy with period being 5, the numerical-based policy, and the optimal policy with respect to the weighting factor ω ( p 0 = 0.2 ,   p 1 = 0.8 ,   q 0 = 0.2 ,   q 1 = 0.5 ,   C e = 1 ).
Figure 6. Comparison of the zero-wait policy, the periodic policy with period being 5, the numerical-based policy, and the optimal policy with respect to the weighting factor ω ( p 0 = 0.2 ,   p 1 = 0.8 ,   q 0 = 0.2 ,   q 1 = 0.5 ,   C e = 1 ).
Entropy 23 00912 g006
Figure 7. Comparison of the zero-wait policy, the periodic policy with period being 5, and the optimal policy with respect to p 1 ( q 0 = 0.2 ,   q 1 = 0.5 ,   ω = 10 ,   C e = 1 ).
Figure 7. Comparison of the zero-wait policy, the periodic policy with period being 5, and the optimal policy with respect to p 1 ( q 0 = 0.2 ,   q 1 = 0.5 ,   ω = 10 ,   C e = 1 ).
Entropy 23 00912 g007
Figure 8. AoI comparison of the zero-wait policy, the periodic policy with period being 5, and the optimal policy with respect to p 1 ( q 0 = 0.2 ,   q 1 = 0.5 ,   ω = 10 ,   C e = 1 ).
Figure 8. AoI comparison of the zero-wait policy, the periodic policy with period being 5, and the optimal policy with respect to p 1 ( q 0 = 0.2 ,   q 1 = 0.5 ,   ω = 10 ,   C e = 1 ).
Entropy 23 00912 g008
Figure 9. Energy consumption comparison of the zero-wait policy, the periodic policy with period being 5, and the optimal policy with respect to p 1 ( q 0 = 0.2 ,   q 1 = 0.5 ,   ω = 10 ,   C e = 1 ).
Figure 9. Energy consumption comparison of the zero-wait policy, the periodic policy with period being 5, and the optimal policy with respect to p 1 ( q 0 = 0.2 ,   q 1 = 0.5 ,   ω = 10 ,   C e = 1 ).
Entropy 23 00912 g009
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Peng, F.; Chen, X.; Wang, X. Channel Quality-Based Optimal Status Update for Information Freshness in Internet of Things. Entropy 2021, 23, 912. https://doi.org/10.3390/e23070912

AMA Style

Peng F, Chen X, Wang X. Channel Quality-Based Optimal Status Update for Information Freshness in Internet of Things. Entropy. 2021; 23(7):912. https://doi.org/10.3390/e23070912

Chicago/Turabian Style

Peng, Fuzhou, Xiang Chen, and Xijun Wang. 2021. "Channel Quality-Based Optimal Status Update for Information Freshness in Internet of Things" Entropy 23, no. 7: 912. https://doi.org/10.3390/e23070912

APA Style

Peng, F., Chen, X., & Wang, X. (2021). Channel Quality-Based Optimal Status Update for Information Freshness in Internet of Things. Entropy, 23(7), 912. https://doi.org/10.3390/e23070912

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop