Abstract
In this paper, the problem of energy-efficient unmanned aerial vehicle (UAV)-assisted multi-access task offloading is investigated. In the studied system, several UAVs are deployed as edge servers to cooperatively aid task executions for several energy-limited computation-scarce terrestrial user equipments (UEs). An expected energy efficiency maximization problem is then formulated to jointly optimize UAV trajectories, UE local central processing unit (CPU) clock speeds, UAV-UE associations, time slot slicing, and UE offloading powers. This optimization is subject to practical constraints, including UAV mobility, local computing capabilities, mixed-integer UAV-UE pairing indicators, time slot division, UE transmit power, UAV computational capacities, and information causality. To tackle the multi-dimensional optimization problem under consideration, the duo-staggered perturbed actor-critic with modular networks (DSPAC-MN) solution in a multi-agent deep reinforcement learning (MADRL) setup, is proposed and tailored, after mapping the original problem into a stochastic (Markov) game. Time complexity and communication overhead are analyzed, while convergence performance is discussed. Compared to representative benchmarks, e.g., multi-agent deep deterministic policy gradient (MADDPG) and multi-agent twin-delayed DDPG (MATD3), the proposed DSPAC-MN is validated to be able to achieve the optimal performance of average energy efficiency, while ensuring 100% safe flights.
| Original language | English |
|---|---|
| Pages (from-to) | 8405-8420 |
| Number of pages | 16 |
| Journal | IEEE Transactions on Communications |
| Volume | 73 |
| Issue number | 9 |
| DOIs | |
| Publication status | Accepted/In press - 2025 |
| Externally published | Yes |
Keywords
- energy efficiency maximization
- Multi-access edge computing (MEC)
- multi-agent deep reinforcement learning (MADRL)
- path planning
- unmanned aerial vehicle (UAV)