TY - JOUR
T1 - Energy-Efficient UAV-Driven Multi-Access Edge Computing
T2 - A Distributed Many-Agent Perspective
AU - Li, Yuanjian
AU - Madhukumar, A. S.
AU - Ernest, Tan Zheng Hui
AU - Zheng, Gan
AU - Saad, Walid
AU - Hamid Aghvami, A.
N1 - Publisher Copyright:
© 1972-2012 IEEE.
PY - 2025
Y1 - 2025
N2 - In this paper, the problem of energy-efficient unmanned aerial vehicle (UAV)-assisted multi-access task offloading is investigated. In the studied system, several UAVs are deployed as edge servers to cooperatively aid task executions for several energy-limited computation-scarce terrestrial user equipments (UEs). An expected energy efficiency maximization problem is then formulated to jointly optimize UAV trajectories, UE local central processing unit (CPU) clock speeds, UAV-UE associations, time slot slicing, and UE offloading powers. This optimization is subject to practical constraints, including UAV mobility, local computing capabilities, mixed-integer UAV-UE pairing indicators, time slot division, UE transmit power, UAV computational capacities, and information causality. To tackle the multi-dimensional optimization problem under consideration, the duo-staggered perturbed actor-critic with modular networks (DSPAC-MN) solution in a multi-agent deep reinforcement learning (MADRL) setup, is proposed and tailored, after mapping the original problem into a stochastic (Markov) game. Time complexity and communication overhead are analyzed, while convergence performance is discussed. Compared to representative benchmarks, e.g., multi-agent deep deterministic policy gradient (MADDPG) and multi-agent twin-delayed DDPG (MATD3), the proposed DSPAC-MN is validated to be able to achieve the optimal performance of average energy efficiency, while ensuring 100% safe flights.
AB - In this paper, the problem of energy-efficient unmanned aerial vehicle (UAV)-assisted multi-access task offloading is investigated. In the studied system, several UAVs are deployed as edge servers to cooperatively aid task executions for several energy-limited computation-scarce terrestrial user equipments (UEs). An expected energy efficiency maximization problem is then formulated to jointly optimize UAV trajectories, UE local central processing unit (CPU) clock speeds, UAV-UE associations, time slot slicing, and UE offloading powers. This optimization is subject to practical constraints, including UAV mobility, local computing capabilities, mixed-integer UAV-UE pairing indicators, time slot division, UE transmit power, UAV computational capacities, and information causality. To tackle the multi-dimensional optimization problem under consideration, the duo-staggered perturbed actor-critic with modular networks (DSPAC-MN) solution in a multi-agent deep reinforcement learning (MADRL) setup, is proposed and tailored, after mapping the original problem into a stochastic (Markov) game. Time complexity and communication overhead are analyzed, while convergence performance is discussed. Compared to representative benchmarks, e.g., multi-agent deep deterministic policy gradient (MADDPG) and multi-agent twin-delayed DDPG (MATD3), the proposed DSPAC-MN is validated to be able to achieve the optimal performance of average energy efficiency, while ensuring 100% safe flights.
KW - energy efficiency maximization
KW - Multi-access edge computing (MEC)
KW - multi-agent deep reinforcement learning (MADRL)
KW - path planning
KW - unmanned aerial vehicle (UAV)
UR - http://www.scopus.com/inward/record.url?scp=105000871413&partnerID=8YFLogxK
U2 - 10.1109/TCOMM.2025.3552746
DO - 10.1109/TCOMM.2025.3552746
M3 - Article
AN - SCOPUS:105000871413
SN - 0090-6778
JO - IEEE Transactions on Communications
JF - IEEE Transactions on Communications
ER -