TY - GEN
T1 - Deep Reinforcement Learning-enabled Dynamic UAV Deployment and Power Control in Multi-UAV Wireless Networks
AU - Bai, Yu
AU - Chang, Zheng
AU - Jantti, Riku
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Using Unmanned Aerial Vehicles (UAVs) as aerial base stations for providing services to ground users has received growing research interest in recent years. The dynamic deployment of UAVs represents a significant research direction within UAV network studies. This paper introduces a highly adaptable UAV wireless network that accounts for the mobility of UAVs and users, the variability in their states, and the tunable transmission power of UAVs. The objective is to maximize energy efficiency while ensuring the minimum number of unserved online users. This dual objective is achieved by jointly optimizing the states, transmission powers, and movement strategies of UAVs. To address the variable state challenges posed by the dynamic environment, user and UAV data is encapsulated within a multi-channel map. A Convolutional Neural Network (CNN) then processes this map to extract key features. The deployment and power control strategy are determined by an agent trained by the Proximal Policy Optimization (PPO)-based Deep Reinforcement Learning (DRL) algorithm. Simulation results demonstrate the effectiveness of the proposed strategy in enhancing energy efficiency and reducing the number of unserved online users.
AB - Using Unmanned Aerial Vehicles (UAVs) as aerial base stations for providing services to ground users has received growing research interest in recent years. The dynamic deployment of UAVs represents a significant research direction within UAV network studies. This paper introduces a highly adaptable UAV wireless network that accounts for the mobility of UAVs and users, the variability in their states, and the tunable transmission power of UAVs. The objective is to maximize energy efficiency while ensuring the minimum number of unserved online users. This dual objective is achieved by jointly optimizing the states, transmission powers, and movement strategies of UAVs. To address the variable state challenges posed by the dynamic environment, user and UAV data is encapsulated within a multi-channel map. A Convolutional Neural Network (CNN) then processes this map to extract key features. The deployment and power control strategy are determined by an agent trained by the Proximal Policy Optimization (PPO)-based Deep Reinforcement Learning (DRL) algorithm. Simulation results demonstrate the effectiveness of the proposed strategy in enhancing energy efficiency and reducing the number of unserved online users.
UR - http://www.scopus.com/inward/record.url?scp=85202840197&partnerID=8YFLogxK
U2 - 10.1109/ICC51166.2024.10622465
DO - 10.1109/ICC51166.2024.10622465
M3 - Conference article in proceedings
AN - SCOPUS:85202840197
T3 - IEEE International Conference on Communications
SP - 1286
EP - 1291
BT - ICC 2024 - IEEE International Conference on Communications
A2 - Valenti, Matthew
A2 - Reed, David
A2 - Torres, Melissa
PB - IEEE
T2 - IEEE International Conference on Communications
Y2 - 9 June 2024 through 13 June 2024
ER -