Next Article in Journal
Application of the Complex Moments for Selection of an Optimal Sensor
Next Article in Special Issue
Power Allocation and Energy Cooperation for UAV-Enabled MmWave Networks: A Multi-Agent Deep Reinforcement Learning Approach
Previous Article in Journal
Quad-Port Multiservice Diversity Antenna for Automotive Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Deep Reinforcement Learning for UAV Trajectory Design Considering Mobile Ground Users

1
School of Information and Communication Engineering, Chungbuk National University, Chungju 28644, Korea
2
Electronics and Telecommunications Research Institute, Daejeon 34129, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(24), 8239; https://doi.org/10.3390/s21248239
Submission received: 9 September 2021 / Revised: 2 December 2021 / Accepted: 6 December 2021 / Published: 9 December 2021
(This article belongs to the Special Issue UAV Control and Communications in 5G and beyond Networks)

Abstract

:
A network composed of unmanned aerial vehicles (UAVs), serving as base stations (UAV-BS network), is emerging as a promising component in next-generation communication systems. In the UAV-BS network, the optimal positioning of a UAV-BS is an essential requirement to establish line-of-sight (LoS) links for ground users. A novel deep Q-network (DQN)-based learning model enabling the optimal deployment of a UAV-BS is proposed. Moreover, without re-learning of the model and the acquisition of the path information of ground users, the proposed model presents the optimal UAV-BS trajectory while ground users move. Specifically, the proposed model optimizes the trajectory of a UAV-BS by maximizing the mean opinion score (MOS) for ground users who move to various paths. Furthermore, the proposed model is highly practical because, instead of the locations of individual mobile users, an average channel power gain is used as an input parameter. The accuracy of the proposed model is validated by comparing the results of the model with those of a mathematical optimization solver.

1. Introduction

Due to the advantages of high mobility and easy deployment, unmanned aerial vehicles (UAVs) are emerging as a major component in various applications such as a mobile access point for military operations, a remote structural safety diagnosis, a quick deployment of communication infrastructure for disaster relief, and agricultural monitoring, etc. Naturally, related research works are continually conducted for the efficient utilization of them. In a mobile network, UAV is considered to play important roles as an aerial user and an aerial base station (BS). As an aerial user, UAV is being adopted in various fields such as Structural Health Monitoring (SHM) [1], disaster relief networks [2], agricultural applications [3], search and rescue (SAR) [4], and aerial ad-hoc networks [5,6], etc. Moreover, a network with UAVs, which serve as aerial BSs (UAV-BSs), is becoming a key component in next-generation mobile communication systems. With the ever-demanding request for high-speed mobile communication, next-generation mobile communication technologies are focusing on the efficient use of wide bandwidth. Accordingly, in addition to installing a larger number of BSs, it is necessary to increase the number of line-of-sight (LoS) links between BSs and mobile users. Since a UAV-BS can be located at a high altitude, it has an advantage in supporting LoS links. Compared to a terrestrial BS, a UAV-BS has few installation restrictions and a low installation cost. Furthermore, a UAV-BS can be readily moved to the vicinity of hotspot areas. Considering these features, a network with UAV-BSs is a promising technology for the next-generation networks.
When a UAV acts as a BS, locating the UAV to a proper position is a critical issue. The location of a UAV-BS in a network largely determines the energy consumption of the UAV-BS, the number of users who can be serviced, and the quality-of-experience (QoE) of the users. Accordingly, a lot of research has been conducted to find the optimal position of a UAV-BS, and the main objective of this research is reducing the energy consumption of UAV-BS, improving service quality for users, and maximizing the coverage of UAV-BS, etc. [7,8,9,10,11,12,13,14,15,16]. In [7,8,9,10], optimization algorithms and mathematical optimization solvers are adopted to obtain the optimal location of a UAV-BS. However, the approach of calculating the optimal position of a UAV-BS through an optimization solver has a considerable computational complexity, because whenever the topology of a network changes, it demands re-calculation. On the other hand, reinforcement learning is very efficient in deriving the optimal UAV-BS location and provides a versatile model applicable to various user distributions [17]. A Q-learning algorithm [18] is one of the popular methods for reinforcement. However, Q-learning has a drawback in that the number of states increases explosively as the number of input variables increases, and its memory usage also increases sharply since it should store all the state-action relations in a table. Accordingly, many related works [11,12,13,14,15,16] adopt Deep Q-Network (DQN) [19], which combines Q-learning with an artificial neural network [20].
In [21,22,23], the authors derive the optimal trajectory and path of a UAV-BS using Q-learning. The objectives are to maximize the sum-rate [21], to maximize the QoE of users [22], and to maximize the number and fairness of users served [23]. In these papers, the altitude of a UAV-BS is fixed, and the results are 2-D trajectories of the UAV-BS. By contrast, in this paper, a DQN model producing 3-D trajectories is proposed where the altitude of a UAV-BS is adjusted according to the density of ground users (GUs). In [11,12,13], the optimal UAV-BS deployment algorithms through DQN are proposed. In [11], a network utility and a tolerable convergence speed are maximized. In [12,13], the number of served aerial nodes and an average user throughput is considered. However, these research works do not take the mobility of users into account. Meanwhile, in [14,15,16], the mobility of users is considered in the optimal trajectory design of a UAV-BS. In [14], an uplink sum-rate is maximized by taking both aerial users and GUs into account. In [15], the QoE of aerial users is maximized. However, in [14,15] the location information of all the users and the UAV-BS is required as input parameters for the proposed DQN learning models, and the results do not contain explicit UAV-BS trajectories following mobile users. In [16], an uplink sum-rate is maximized using signal strength as an input parameter for a DQN, and the trajectory of a UAV-BS is presented. However, a simple user mobility model is considered, where GUs move to a specific position only once, and trajectory results for various paths are insufficient. By contrast, the output trajectory of the proposed DQN model in this paper dynamically follows mobile GUs, which move various courses. Moreover, the results of these papers do not show a clear 3-D trajectory of the UAV according to various movement paths of GUs.
In this paper, a UAV-BS trajectory design algorithm, which maximizes QoE considering mobile GUs, is proposed. The contribution of this work is summarized as follows:
  • The proposed DQN model exploits an average channel power gain information rather than individual GU position information, which greatly reduces the size of input parameter and computational complexity.
  • Reflecting the density of GUs, the adjustment of UAV-BS altitude is enabled. This leads to 3-D trajectory design according to diverse moving patterns of GUs.
  • The proposed DQN model learns from a static GUs distribution, then the derived model can be applied to mobile GU scenarios in which the proposed model requires neither the user mobility information nor re-learning for the moving GUs.
  • The accuracy of the proposed model is validated by comparing the result of the proposed model with a mathematical optimization solver [24].
Note that applying the proposed DQN model, which is trained in a static GUs distribution, to mobile GU scenarios itself is a great advantage. Because a training a DQN model in a static GU distribution is much easier than training in a mobile GU distribution. In the training of a DQN model with a mobile GUs distribution, the moving pattern of the GUs can be very diverse and the optimal UAV-BS position should be updated in a real-time reflecting the moving GUs.

2. System Model

A single UAV-BS denoted as K and a number of N GUs are considered. It is assumed that the UAV-BS communicates with the GUs using time slots of equal length. It is also assumed that the locations of the UAV-BS and the GUs do not change during the time slot duration. Accordingly, 3-D coordinates of the UAV-BS K at time t is ( x K ( t ) ,   y K ( t ) ,   h K ( t ) ) , and the GUs are assumed to be on the ground with zero height. The coordinates of the GUs are ( x i ( t ) ,   y i ( t ) ,   0 ) ,   i = 1 , , N . The distance between the UAV-BS and GU i at time t is expressed as d i ( t ) = ( x K ( t ) x i ( t ) ) 2 + ( y K ( t ) y i ( t ) ) 2 + h K 2 ( t ) ,   i = 1 ,   ,   N   . For the sake of clarity, the notations and the associated descriptions are provided in Table 1.

2.1. Air to Ground Model

The air to ground model considers both LoS and non-LoS (NLoS) characteristics, and the probabilities of connecting LoS and NLoS links are as follows:
P L O S ( θ i ) = 1 1 + a exp ( b [ θ i a ] )  
P N L O S ( θ i ) = 1 P L O S ( θ i ) ,
where θ i is an elevation angle between the UAV-BS and GU i , and a and b are constants to be determined according to the surrounding environment (urban, sub-urban, rural, …). It is assumed that the bandwidth and the transmission power of the UAV-BS are equally allocated to all the GUs. Hence, the bandwidth B i = B / N and the transmission power p i = P / N are allocated to GU i , where B and P denote the total bandwidth and the total transmission power, respectively. Then, the received SNR Γ i ( t ) and the transmission rate T i ( t ) of the GU i at time t are expressed as follows:
Γ i ( t ) = p i g i ( t ) B i N 0 ,
T i ( t ) = B i log 2 ( 1 + Γ i ( t ) ) ,
where N 0 is the noise power spectral density and g i ( t ) is the channel gain between the UAV-BS and GU i , which is given by [15]:
g i ( t ) = K 0 1 d i α ( t ) [ P L O S μ L O S + P N L O S μ N L O S ] 1 ,
where K 0 = ( 4 π f c c ) 2 , α is a path loss exponent, and μ L o s and μ N L o s are attenuation factors for LoS and NLoS, respectively.
The UAV-BS receives feedback information of the channel power gain from each GU. These received gains are averaged to an average received channel power gain. Therefore, it is possible to simplify the model by reducing the dimension of the input parameter.

2.2. QoE Model

QoE is the quality of service experienced by GUs, and mean opinion score (MOS) is a representative metric for QoE. We adopt the MOS model applicable in the TCP protocol proposed in [25]. The simplified MOS model for GU i is as follows [15]:
MOS i ( t )   =   C 1 ln [ d ( T i ( t ) ) ] + C 2
where C 1 ,   C 2 are given constants, d ( T i ( t ) ) is a delay related to the transmission rate for GU i , which is expressed as [26]
d ( T i ( t ) ) = T L / ( T i ( t ) )  
where T L is traffic load.

2.3. User Mobility Model

It is assumed that the GUs move randomly within a certain radius around a moving center point while this moving center point moves along a predefined path. The radius may vary over time, and the change of the radius results in the variation of the density of the GUs. The GUs are uniform randomly distributed within the radius. Figure 1 shows the user mobility model schematically.

3. Proposed Algorithm

The proposed trajectory design model optimizes not only the horizontal coordinates of the UAV-BS but the altitude of it as well. The model is learned through a DQN by maximizing the MOS of the GUs.

3.1. Problem Formulation

In this paper, the goal of the algorithm is to maximize QoE by considering mobile GUs. Hence, the problem formulation maximizes the MOS and can be expressed as:
max x K ( t ) , y K ( t ) , h K ( t ) i = 1 N MOS i ( t ) x min x K ( t ) x max , y min y K ( t ) y max , h min h K ( t ) h max ,
where the minimum and maximum values of x ,   y , and h are grid sizes, indicating the area in which the UAV can fly. We solve this problem using our proposed DQN model and the optimization solver using the Broyden–Fletcher–Goldfarb–Shanno (BFGS) [24] algorithm, and then we compare the results of the two methods.

3.2. MDP

Action: a set of actions includes the 3-D movements of the UAV-BS. Accordingly, the UAV-BS has horizontal actions forward (F), backward (B), right (R), left (L), and vertical actions up (U), down (D), and finally staying in place (S). This action set considers a total seven actions, which are expressed as A = { F , B , R , L , U , D , S } .
When each action is selected, the UAV-BS moves along the selected direction by a predefined distance δ m . However, when the received average channel power gain of the GUs is lower than a threshold τ , which means the UAV-BS is far from the optimal position, the UAV-BS moves with a larger step size Δ m ( > δ m ) . This mechanism allows the UAV-BS to move to the optimal position quickly when the UAV-BS is initially located far away from the optimal position. In addition, after moving to the optimal position, the staying action prevents the UAV-BS from unnecessary maneuvering.
State: from the above action set, three flying directions, i.e., F-B, R-L, and U-D, can be considered. In describing a state in the proposed model, three parameters constitute a state. Specifically, the differences of the average received channel power gain in F-B, R-L, and U-D directions constitute a state vector. For instance, if F action is selected and the UAV-BS moves to a new position, the new average channel power gain is subtracted by the previous value, and F-B direction element of the state vector is updated to this value. The state vector at time t is expressed as s t   = [ Δ F B , Δ R L , Δ U D ] , where Δ F B ,   Δ R L , and Δ U D are the difference in received channel power gain in the F-B, R-L, and U-D directions, respectively. For instance, let a UAV-BS move in order of R, F, R, and B from t 4 to t and the average received channel power gains be a ,   b ,   c ,   d , and e , as shown in Figure 2. In this case, the state vector at each time step is given in Table 2.
Reward: The reward at time slot t is expressed as r t . The optimal UAV-BS position maximizes the sum of the GUs’ MOS. Accordingly, it is quite a natural and general approach to allocate a positive reward α to an action of increasing MOS and to allocate a negative reward α to an action of decreasing MOS. Moreover, to prevent the oscillation of the UAV-BS position, a small positive reward β is allocated to an action of retaining current MOS. In the experimental results section, Figure 3 of training process confirms that the reward function operates successfully.

3.3. Deep Q-Network (DQN) Algorithm

An ϵ greedy approach is adopted because exploration for learning is necessary. At the start of the algorithm, the probability of exploration is increased by setting ϵ to 1. Subsequently, the probability ϵ is reduced by multiplying ϵ decay at every time step. The UAV-BS moves by δ m in the selected direction of a t . If the randomly chosen action by the ϵ -probability goes out of the area grid, the UAV-BS randomly takes another action. The DQN algorithm for the UAV-BS trajectory model learning is shown in the following Algorithm 1.
Algorithm 1. DQN algorithm for UAV-BS trajectory
• Initialize the replay memory D
• Initialize action-value function Q with random weights
• Initialize target-value function Q’ with random weights
• Initialize the position of N GUs with R radius.
• Set probability ϵ = 1, ϵ min = 0.1, ϵ decay = 0.99997
1: for episode = 1 ,   ,   M do
2:          Initialize the position of the UAV-BS
3:          for  t = 1 ,   ,   T do
4:                if  ϵ >   ϵ min
5:                           ϵ   =   ϵ × ϵ decay
6:                end if
7:                Select a random action a t with probability
8:                while the UAV-BS position goes out of the grid
9:                          Select other action except a t
10:            end while
11:            otherwise select a t = arg   max a Q ( s t , a )
12:            Execute action a t and observe reward r t and state S t + 1
13:            Store transition ( S t , a t , r t , S t + 1 ) in D
14:            Sample random mini-batch of transitions from D
15:            Perform a gradient descent to update action-value function Q
16:            Every episode update target-value function Q’
17:        end for
18: end for

3.4. Algorithm Complexity

The complexity of approximating Q function of DQN is affected by the number of states and can be expressed as O ( | S | 2 | A | ) [11]. | S | and | A | represent the numbers of states and actions, respectively. In the proposed model, the number of components in the state vector is fixed to three regardless of the number of the GUs because the average channel power gain is adopted as an input parameter. Therefore, even if the number of GUs increases, the proposed model has an advantage in terms of computational complexity.

4. Experimental Results

For the performance analysis, 25 GUs are considered in an area where the grid size is 300 m × 300 m, and the maximum altitude of the UAV-BS is 50 m. The parameter settings for the DQN learning and the experiment parameters are summarized in Table 3 and Table 4, respectively. In determining predefined movement distance δ m , the average UAV-BS altitude 30 m and the average GUs movement per a time slot 1 m are considered, and even when δ m is reduced to 1 m, MOS gain is not observed, hence, δ m = 5 is determined. Initially, the model learns in an environment where the GUs are fixed, and then this learned model is applied to the moving GUs. Note that the model can be trained in both the environments where the GUs are fixed or the GUs move. Since the performance difference is negligible, the model trained with the fixed GUs is preferred in this paper. In the learning stage, 25 GUs are located around the center of an area with 50 m group radius, and it takes 300 episodes with a random UAV-BS initial position in training this model, which has five layers and a rectified linear unit (ReLU) activation function. At the early part of the learning, MOS is about 30, and after the learning, MOS reaches above 50 as shown in Figure 3. In the execution stage, it is assumed that the GUs move between episodes, i.e., no movement within a single episode. At the beginning of the experiment, the position of the UAV-BS is randomly chosen. The position of the UAV-BS and the MOS of GUs are measured by varying the group radius and the path of GUs. In addition, the experimental results are compared with those obtained through the solver of the Python scipy package implemented based on the BFGS algorithm. The solver needs the exact positions of the UAV-BS and all the GUs, and it finds the position of the UAV-BS by locally maximizing the MOS of the GUs. Hence, in order to confirm the optimality of the results of the solver, more than four different initial points are fed to the solver, resulting in the same output of the solver.
Table 5 is a comparison between the proposed algorithm and BFGS. The time complexity of DQN is O ( | S | 2 | A | ) , and the algorithm proposed in this paper has a fixed number of inputs (states). Also, this is the time complexity calculated during the training process. When a trained model is applied in UVA-BS network for execution the time complexity is low O ( | A | ) . On the other hand, in the case of BFGS, the time complexity O ( n 2 ) increases with the number of GUs because the location information of all GUs should be received and processed. Moreover, BFGS requires the exact position of each GU as input data, while the proposed model requires the average received channel power gain. The input data requirement for BFGS is quire impractical because it assumes that all the GUs are equipped with GPS and all the GUs’ position are reported to the UAV-BS in a real time. Considering the time complexities of the two methods, the execution time of BFGS is expected to increase sharply as the number of GUs increases, while the proposed algorithm will maintain its execution time even with increasing number of GUs. Moreover, the output of BFGS is the optimal position of the UAV-BS, while the output of the proposed algorithm is the optimal direction to reach the optimal position. Hence, the proposed algorithm requires some iteration before the accumulated optimal directions guide the UAV-BS to the optimal position. This intuition is very well matched with the measured execution time shown in Figure 4. When the number of GUs is small the execution time of BFGS smaller than the proposed algorithm because the proposed algorithm requires some iteration; however, as the number of GUs increases, the execution time of BFGS increases sharply, while the proposed algorithm maintains its execution time regardless of the number of GUs.
Figure 5, Figure 6 and Figure 7 show the optimal deployment of the UAV-BS and the measured MOS over a single episode. Figure 5 shows the results of an experiment in which group radius is fixed at 50 m. In this figure, even though randomly selected initial points are adopted, the final positions of the learning model result in the nearly same position. Moreover, these final positions have little error compared to the positions obtained by the solver. Moreover, compared with Scipy-BFGS, the MOS gaps between the two methods are negligible. It shows that the proposed model with the predefined moving distance, discrete action setting, and different initial points reaches the same optimal result in terms of both position and performance. Figure 5b shows the LoS probability calculated through Equation (1) at randomly selected initial positions and final positions. When the position of the UAV-BS changes, the elevation angle between the UAV-BS and the GUs changes, so the LoS probability changes. Figure 5b shows the advantage of moving the UAV-BS to the optimal position. As shown in this figure, the probabilities of establishing LoS links with GUs are very low with initial points like 0.05, 0.1, 0.2, 0.5; however, at the final points, these probabilities become higher than 0.9, which results in the improved channel quality between the UAV-BS and the GUs.
Figure 6 and Figure 7 show the optimal UAV-BS altitude and MOS with varying group radius, respectively. The initial group radius is 10 m and increases by 10 m in each episode. As the group radius increases, the NLoS probability for the GUs located near the boundary of group increases. This leads to an increment of the average of NLOS probability. Hence, the UAV-BS starts to decrease the NLoS probability by increasing its altitude. Accordingly, as shown in Figure 6, the altitude of the UAV-BS is linearly proportional to the group radius, approximately 6 m (altitude) per 10 m (radius). However, too high an altitude reduces the MOS of the GUs. Therefore, it is important to find the proper altitude of the UAV-BS.
There are very small discrepancies between the UAV-BS positions out of the proposed model and those of the mathematical solver. Moreover, the difference in terms of MOS is negligible, as shown in Figure 7. This means that the position difference has a very slight effect to the QoE of the GUs. In Figure 7, MOS decreases as the group radius increases. As mentioned above, the wide group radius leads to the high altitude of the UAV-BS, and it results in the decreased received channel power gain and the decreased MOS.
The results shown in Figure 8 is very promising and validate the effectiveness of the proposed model. This figure shows the optimal trajectories and the associated MOS curves of the proposed model applied to various paths of the mobile GUs. Note that the model learns in a static environment where the GUs are fixed, then, without the acquisition of the path information and re-learning, this model is applied to the mobile environments where the GUs move randomly around the moving center point. As the GUs move episode by episode, the radius of moving GUs is randomly changed from 25 m to 50 m. In (a) of Figure 8, the GUs move in one direction, and in cases of (c) and (e), the GUs show more dynamic moving patterns. To confirm that the results are not accidental, the experiments are repeated over 50 times, then averaged. In each repetition, the movement of individual GU is randomized with a new random seed. As we can see in Figure 8, even though the situation is so adverse that the GUs’ position information is unavailable and the GUs move with varying group radius, the proposed model successfully locates the UAV-BS at the optimal position. In addition, if the model learns in a dynamic environment where the GUs move randomly episode by episode, the output trajectories are nearly same with those shown in Figure 8. Considering that the proposed DQN model are learned with simplified parameters, and optimal actions selected by the UAV-BS are predefined distance values consisting of seven discrete directions, the proposed algorithm and the optimal position derived by BFGS are very close, as shown in the resulting graph. Moreover, in terms of MOS, the outputs of the proposed model very well matched with those of the solver.

5. Conclusions

In this paper, a novel DQN model for an optimal deployment and trajectory design of a UAV-BS is proposed. This model uses only the average channel power gain without accurate location information of GUs. It is confirmed that the proposed model locates UAV-BS where MOS is maximized. Experimental results show that the altitude of UAV-BS increases as the group radius increases. In addition, they demonstrate that 3D trajectory design of UAV-BS is possible using the DQN-model where the model learns in a static environment, and then this model is applied to mobile environments without re-learning.

Author Contributions

W.L., Y.J., T.K. and Y.-I.K. conceived and designed the experiments; W.L. and Y.J. performed the network simulation; W.L, analyzed the data; Y.-I.K. acquired funding; W.L. and T.K. wrote the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Institute of Civil Military Technology Cooperation funded by the Defense Acquisition Program Administration and Ministry of Trade, Industry and Energy of Korean government under Grant 18-CM-SW-19.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bacco, M.; Barsocchi, P.; Cassará, P.; Germanese, D.; Gotta, A.; Leone, G.-R.; Moroni, D.; Pascali, M.-A.; Tampucci, M. Monitoring ancient buildings: Real deployment of an IoT system enhanced by UAVs and virtual reality. IEEE Access 2020, 8, 50131–50148. [Google Scholar] [CrossRef]
  2. Wang, B.; Sun, Y.; Liu, D.; Nguyen, H.-M.; Duong, T.-Q. Social-aware UAV-assisted mobile crowd sensing in stochastic and dynamic environments for disaster relief networks. IEEE Trans. Veh. Technol. 2020, 69, 1070–1074. [Google Scholar] [CrossRef]
  3. Costa, F.-G.; Ueyama, J.; Braun, T.; Pessin, G.; Osorio, F.-S.; Vargas, P.-A. The use of unmanned aerial vehicles and wireless sensor network in agricultural applications. In Proceedings of the 2012 IEEE International Geoscience and Remote Sensing Symposium, Munich, Germany, 22–27 July 2012; pp. 5045–5048. [Google Scholar]
  4. Hayat, S.; Yanmaz, E.; Brown, T.-X.; Bettstetter, C. Multi-objective UAV path planning for search and rescue. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 5569–5574. [Google Scholar] [CrossRef]
  5. Kumar, K.; Kumar, S.; Kaiwartya, O.; Kashyap, P.K.; Lloret, J.; Song, H. Drone assisted Flying Ad-Hoc Networks: Mobility and Service oriented modeling using Neuro-fuzzy. Ad Hoc Netw. 2020, 106, 102242. [Google Scholar] [CrossRef]
  6. Kumar, K.; Kumar, S.; Kaiwartya, O.; Sikandar, A.; Kharel, R.; Mauri, J.L. Internet of Unmanned Aerial Vehicles: QoS Provisioning in Aerial Ad-Hoc Networks. Sensors 2020, 20, 3160. [Google Scholar] [CrossRef] [PubMed]
  7. Cicek, C.T.; Gultekin, H.; Tavli, B.; Yanikomeroglu, H. UAV Base Station Location Optimization for Next Generation Wireless Networks: Overview and Future Research Directions. In Proceedings of the 2019 1st International Conference on Unmanned Vehicle Systems-Oman (UVS) 2019, Muscat, Oman, 5–7 February 2019; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  8. Lai, C.; Chen, C.; Wang, L. On-Demand Density-Aware UAV Base Station 3D Placement for Arbitrarily Distributed Users With Guaranteed Data Rates. IEEE Wirel. Commun. Lett. 2019, 8, 913–916. [Google Scholar] [CrossRef] [Green Version]
  9. Cherif, N.; Jaafar, W.; Yanikomeroglu, H.; Yongacoglu, A. On the Optimal 3D Placement of a UAV Base Station for Maximal Coverage of UAV Users. In Proceedings of the GLOBECOM 2020—2020 IEEE Global Communications Conference, Taipei, Taiwan, 7–11 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
  10. Alzenad, M.; El-Keyi, A.; Lagum, F.; Yanikomeroglu, H. 3-D Placement of an Unmanned Aerial Vehicle Base Station (UAV-BS) for Energy-Efficient Maximal Coverage. IEEE Wirel. Commun. Lett. 2017, 6, 434–437. [Google Scholar] [CrossRef] [Green Version]
  11. Li, X.; Wang, Q.; Liu, J.; Zhang, W. 3D Deployment with machine learning and system performance analysis of UAV-enabled networks. In Proceedings of the 2020 IEEE/CIC International Conference on Communications in China (ICCC), Chongqing, China, 9–11 August 2020; pp. 554–559. [Google Scholar] [CrossRef]
  12. Hou, M.-C.; Deng, D.-J.; Wu, C.-L. Optimum aerial base station deployment for UAV networks: A reinforcement learning approach. In Proceedings of the 019 IEEE Globecom Workshops (GC Wkshps), Waikoloa, HI, USA, 9–13 December 2019; pp. 1–6. [Google Scholar] [CrossRef]
  13. Saxena, V.; Jaldén, J.; Klessig, H. Optimal UAV base station trajectories using flow-level models for reinforcement learning. IEEE Trans. Cogn. Commun. Netw. 2019, 5, 1101–1112. [Google Scholar] [CrossRef]
  14. Hao, G.; Ni, W.; Tian, H.; Cao, L. Mobility-aware trajectory design for aerial base station using deep reinforcement learning. In Proceedings of the 2020 International Conference on Wireless Communications and Signal Processing (WCSP), Nanjing, China, 21–23 October 2020; pp. 1131–1136. [Google Scholar] [CrossRef]
  15. Liu, X.; Liu, Y.; Chen, Y. Reinforcement learning in multiple-UAV networks: Deployment and movement design. IEEE Trans. Veh. Technol. 2019, 68, 8036–8049. [Google Scholar] [CrossRef] [Green Version]
  16. Yin, S.; Zhao, S.; Zhao, Y.; Yu, F.-R. Intelligent trajectory design in UAV-aided communications with reinforcement learning. IEEE Trans. Veh. Technol. 2019, 68, 8227–8231. [Google Scholar] [CrossRef]
  17. Wu, J.; Yu, P.; Feng, L.; Zhou, F.; Li, W.; Qiu, X. 3D aerial base station position planning based on deep Q-network for capacity enhancement. In Proceedings of the 2019 IFIP/IEEE Symposium on Integrated Network and Service Management (IM), Washington, DC, USA, 8–12 April 2019; pp. 482–487. [Google Scholar]
  18. Watkins, J.-C.-H.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  19. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Hassabis, D. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
  20. McClelland, J.L.; Rumelhart, D.E.; Group, T.P.R. Parallel Distributed Processing: Explorations in the Microstructure of Cognition; MIT Press: Cambridge, MA, USA, 1986. [Google Scholar]
  21. Bayerlein, H.; De Kerret, P.; Gesbert, D. Trajectory optimization for autonomous flying base station via reinforcement learning. In Proceedings of the IEEE 19th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Kalamata, Greece, 25–28 June 2018; pp. 1–5. [Google Scholar]
  22. Colonnese, S.; Cuomo, F.; Pagliari, G.; Chiaraviglio, L. Q-SQUARE: A Q-learning approach to provide a QoE aware UAV flight path in cellular networks. Ad Hoc Netw. 2019, 91, 101872. [Google Scholar] [CrossRef]
  23. Abeywickrama, H.V.; He, Y.; Dutkiewicz, E.; Jayawickrama, B.A.; Mueck, M. A Reinforcement Learning Approach for Fair User Coverage Using UAV Mounted Base Stations Under Energy Constraints. IEEE Open J. Veh. Technol. 2020, 1, 67–81. [Google Scholar] [CrossRef]
  24. Wright, S.; Nocedal, J. Numerical Optimization; Springer: New York, NY, USA, 1999. [Google Scholar]
  25. Rugelj, M.; Sedlar, U.; Volk, M.; Sterle, J.; Hajdinjak, M.; Kos, A. Novel cross-layer QoE-aware radio resource allocation algorithms in multiuser OFDMA systems. IEEE Commun. Mag. 2014, 62, 3196–3208. [Google Scholar] [CrossRef]
  26. Cui, J.; Liu, Y.; Ding, Z.; Fan, P.; Nallanathan, A. QoE-Based Resource Allocation for Multi-Cell NOMA Networks. IEEE Trans. Wirel. Commun. 2018, 17, 6160–6176. [Google Scholar] [CrossRef]
Figure 1. User mobility model with varying user density and random movement around moving center point.
Figure 1. User mobility model with varying user density and random movement around moving center point.
Sensors 21 08239 g001
Figure 2. Example of unmanned aerial vehicles serving as base stations (UAV-BS) movement. a ,   b ,   c ,   d , and e mean the average received channel power gains from t 4 to t .
Figure 2. Example of unmanned aerial vehicles serving as base stations (UAV-BS) movement. a ,   b ,   c ,   d , and e mean the average received channel power gains from t 4 to t .
Sensors 21 08239 g002
Figure 3. Training process of the deep Q-network (DQN) model.
Figure 3. Training process of the deep Q-network (DQN) model.
Sensors 21 08239 g003
Figure 4. Execution times of two methods with increasing GUs.
Figure 4. Execution times of two methods with increasing GUs.
Sensors 21 08239 g004
Figure 5. UAV-BS movement and line-of-sight (LoS) probability from random starting positions to the optimal position without considering GU mobility.
Figure 5. UAV-BS movement and line-of-sight (LoS) probability from random starting positions to the optimal position without considering GU mobility.
Sensors 21 08239 g005
Figure 6. The change of UAV-BS position by varying group radius.
Figure 6. The change of UAV-BS position by varying group radius.
Sensors 21 08239 g006
Figure 7. Mean opinion score (MOS) and UAV-BS altitude with varying group radius.
Figure 7. Mean opinion score (MOS) and UAV-BS altitude with varying group radius.
Sensors 21 08239 g007
Figure 8. UAV-BS trajectory and MOS with GUs moving in various paths.
Figure 8. UAV-BS trajectory and MOS with GUs moving in various paths.
Sensors 21 08239 g008aSensors 21 08239 g008b
Table 1. Mathematical notations and descriptions.
Table 1. Mathematical notations and descriptions.
NotationsDescription
θ i Elevation   angle   between   the   unmanned   aerial   vehicles   base   station   ( UAV - BS )   and   ground   user   ( GU )   i
a ,   b Environmental parameters
B ,   B i Total   Bandwidth / Allocated   to   GU   i
P ,   p i Total   transmission   power / Allocated   to   GU   i
Γ i ( t ) Received   signal   to   noise   ratio   ( SNR )   of   GU   i   at   time   slot   t
T i ( t ) Transmission   rate   of   GU   i   at   time   slot   t
N 0 Noise power spectral
g i ( t ) Channel   gain   between   the   UAV - BS   and   GU   i
α Path loss exponent
μ L o s ,   μ N L o s Attenuation factors for line of sight (LoS) and non-LoS (NLoS)
MOS i ( t ) MOS   of   GU   i   at   time   slot   t
d ( T i ( t ) ) Delay   related   to   the   transmission   rate   for   GU   i
T L Traffic load
Table 2. Input vector over time and state.
Table 2. Input vector over time and state.
TimeState
s t = [ Δ F B ,   Δ R L ,   Δ U D ]
t 4 [ 0 ,   0 ,   0 ]
t 3 [ 0 ,   b a ,   0 ]
t 2 [ c b ,   b a ,   0 ]
t 1 [ c b ,   d c ,   0 ]
t [ e d ,   d c , 0 ]
Table 3. DQN learning parameter settings.
Table 3. DQN learning parameter settings.
ParameterValue
Batch size64
Learning rate0.001
Size of replay memory5000
Number of hidden layers 2
Number of neurons in each hidden layer48
Type of activation functionRectified linear unit (ReLU)
Table 4. Experiment parameter settings.
Table 4. Experiment parameter settings.
Time ParameterValue
Number   of   users   N 25
Group   radius   R 10–50 m
Carrier   frequency   f 2 GHz
Transmit power20 dBm
Bandwidth   B 1 MHz
T L 8,000,000 bits
a ,   b 9.61, 0.16
C 1 ,   C 2 1.120, 4.6746
Path   loss   exponent   α 2
μ L o S 3 dB
μ N L o S 23 dB
Movement   distance   δ m   ,   Δ m 5, 10 m
Channel   power   gain   of   threshold   τ −100 dBm
α ,   β 10, 1
Table 5. Comparison between the proposed algorithm and Broyden–Fletcher–Goldfarb–Shanno (BFGS).
Table 5. Comparison between the proposed algorithm and Broyden–Fletcher–Goldfarb–Shanno (BFGS).
BFGSProposed Algorithm
Time complexity High   ( O ( n 2 ) ) Training :   High   ( O ( | S | 2 | A | ) )
Prediction :   O ( | A | )
InputExact positions of the UAV-BS and all the GUsDifferences of the average received channel power gain
OutputOptimal position(coordinate)Optimal action(direction)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, W.; Jeon, Y.; Kim, T.; Kim, Y.-I. Deep Reinforcement Learning for UAV Trajectory Design Considering Mobile Ground Users. Sensors 2021, 21, 8239. https://doi.org/10.3390/s21248239

AMA Style

Lee W, Jeon Y, Kim T, Kim Y-I. Deep Reinforcement Learning for UAV Trajectory Design Considering Mobile Ground Users. Sensors. 2021; 21(24):8239. https://doi.org/10.3390/s21248239

Chicago/Turabian Style

Lee, Wonseok, Young Jeon, Taejoon Kim, and Young-Il Kim. 2021. "Deep Reinforcement Learning for UAV Trajectory Design Considering Mobile Ground Users" Sensors 21, no. 24: 8239. https://doi.org/10.3390/s21248239

APA Style

Lee, W., Jeon, Y., Kim, T., & Kim, Y. -I. (2021). Deep Reinforcement Learning for UAV Trajectory Design Considering Mobile Ground Users. Sensors, 21(24), 8239. https://doi.org/10.3390/s21248239

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop