TY - GEN
T1 - Traffic Steering for Cellular-Enabled UAVs: A Federated Deep Reinforcement Learning Approach
AU - Hellaoui, Hamed
AU - Yang, Bin
AU - Taleb, Tarik
AU - Manner, Jukka
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - This paper investigates the fundamental traffic steering issue for cellular-enabled unmanned aerial vehicles (UAVs), where each UAV needs to select one from different Mobile Network Operators (MNOs) to steer its traffic for improving the Quality-of-Service (QoS). To this end, we first formulate the issue as an optimization problem aiming to minimize the maximum outage probabilities of the UAVs. This problem is non-convex and non-linear, which is generally difficult to be solved. We propose a solution based on the framework of deep reinforcement learning (DRL) to solve it, in which we define the environment and the agent elements. Furthermore, to avoid sharing the learned experiences by the UAV in this solution, we further propose a federated deep reinforcement learning (FDRL)-based solution. Specifically, each UAV serves as a distributed agent to train separate model, and is then communicated to a special agent (dubbed coordinator) to aggregate all training models. Moreover, to optimize the aggregation process, we also introduce a FDRL with DRL-based aggregation (DRL2A) approach, in which the coordinator implements a DRL algorithm to learn optimal parameters of the aggregation. We consider deep Q-learning (DQN) algorithm for the distributed agents and Advantage Actor-Critic (A2C) for the coordinator. Simulation results are presented to validate the effectiveness of the proposed approach.
AB - This paper investigates the fundamental traffic steering issue for cellular-enabled unmanned aerial vehicles (UAVs), where each UAV needs to select one from different Mobile Network Operators (MNOs) to steer its traffic for improving the Quality-of-Service (QoS). To this end, we first formulate the issue as an optimization problem aiming to minimize the maximum outage probabilities of the UAVs. This problem is non-convex and non-linear, which is generally difficult to be solved. We propose a solution based on the framework of deep reinforcement learning (DRL) to solve it, in which we define the environment and the agent elements. Furthermore, to avoid sharing the learned experiences by the UAV in this solution, we further propose a federated deep reinforcement learning (FDRL)-based solution. Specifically, each UAV serves as a distributed agent to train separate model, and is then communicated to a special agent (dubbed coordinator) to aggregate all training models. Moreover, to optimize the aggregation process, we also introduce a FDRL with DRL-based aggregation (DRL2A) approach, in which the coordinator implements a DRL algorithm to learn optimal parameters of the aggregation. We consider deep Q-learning (DQN) algorithm for the distributed agents and Advantage Actor-Critic (A2C) for the coordinator. Simulation results are presented to validate the effectiveness of the proposed approach.
KW - Cellular Networks
KW - Connection Steering
KW - Deep Reinforcement Learning (DRL)
KW - FDRL with DRL-based Aggregation (DRL2A)
KW - Federated Deep Reinforcement Learning (FDRL)
KW - Unmanned Aerial Vehicles (UAVs)
UR - http://www.scopus.com/inward/record.url?scp=85178281690&partnerID=8YFLogxK
U2 - 10.1109/ICC45041.2023.10279441
DO - 10.1109/ICC45041.2023.10279441
M3 - Conference article in proceedings
AN - SCOPUS:85178281690
T3 - IEEE International Conference on Communications
SP - 6230
EP - 6235
BT - ICC 2023 - IEEE International Conference on Communications
A2 - Zorzi, Michele
A2 - Tao, Meixia
A2 - Saad, Walid
PB - IEEE
T2 - IEEE International Conference on Communications
Y2 - 28 May 2023 through 1 June 2023
ER -