Abstract
In this paper, the problem of energy-efficient unmanned aerial vehicle (UAV)-assisted multi-access task offloading is investigated. In the studied system, several UAVs are deployed as edge servers to cooperatively aid task executions for several energy-limited computation-scarce terrestrial user equipments (UEs). An expected energy efficiency maximization problem is then formulated to jointly optimize UAV trajectories, UE local central processing unit (CPU) clock speeds, UAV-UE associations, time slot slicing, and UE offloading powers. This optimization is subject to practical constraints, including UAV mobility, local computing capabilities, mixed-integer UAV-UE pairing indicators, time slot division, UE transmit power, UAV computational capacities, and information causality. To tackle the multi-dimensional optimization problem under consideration, the duo-staggered perturbed actor-critic with modular networks (DSPAC-MN) solution in a multi-agent deep reinforcement learning (MADRL) setup, is proposed and tailored, after mapping the original problem into a stochastic (Markov) game. Time complexity and communication overhead are analyzed, while convergence performance is discussed. Compared to representative benchmarks, e.g., multi-agent deep deterministic policy gradient (MADDPG) and multi-agent twin-delayed DDPG (MATD3), the proposed DSPAC-MN is validated to be able to achieve the optimal performance of average energy efficiency, while ensuring 100% safe flights.
Original language | English |
---|---|
Journal | IEEE Transactions on Communications |
DOIs | |
Publication status | Accepted/In press - Mar 2025 |
Externally published | Yes |
Keywords
- Multi-access edge computing (MEC)
- Unmanned aerial vehicle (UAV)
- Multi-agent deep reinforcement learning (MADRL)
- Energy efficiency maximization
- Path planning