TY - JOUR
T1 - Augmented Deep Reinforcement Learning for Online Energy Minimization of Wireless Powered Mobile Edge Computing
AU - Chen, Xiaojing
AU - Dai, Weiheng
AU - Ni, Wei
AU - Wang, Xin
AU - Zhang, Shunqing
AU - Xu, Shugong
AU - Sun, Yanzan
N1 - Publisher Copyright:
© 1972-2012 IEEE.
PY - 2023/5/1
Y1 - 2023/5/1
N2 - Mobile edge computing (MEC) offers an opportunity for devices relying on wireless power transfer (WPT), to accomplish computationally demanding tasks. Such WPT-powered MEC systems have yet to be optimized for long-term efficiency, due to random and changing task demands and wireless channel states of the devices. This paper presents an augmented two-staged deep Q-network (DQN), referred to as 'TS-DQN,' for online optimization of WPT-powered MEC systems, where the WPT, offloading schedule, channel allocation, and the CPU configurations of the edge server and devices are jointly optimized to minimize the long-term average energy requirement of the systems. The key idea is to design a DQN for learning the channel allocation and task admission, while the WPT, offloading time and CPU configurations are efficiently optimized to precisely evaluate the reward of the DQN and substantially reduce its action space. Another important aspect is that a new action generation method is developed to expand and diversify the actions of the DQN, further accelerating its convergence. As validated by simulations, the proposed TS-DQN is much more energy efficient and converges much faster, than its potential alternative directly using the state-of-the-art Deep Deterministic Policy Gradient algorithm to learn all decision variables.
AB - Mobile edge computing (MEC) offers an opportunity for devices relying on wireless power transfer (WPT), to accomplish computationally demanding tasks. Such WPT-powered MEC systems have yet to be optimized for long-term efficiency, due to random and changing task demands and wireless channel states of the devices. This paper presents an augmented two-staged deep Q-network (DQN), referred to as 'TS-DQN,' for online optimization of WPT-powered MEC systems, where the WPT, offloading schedule, channel allocation, and the CPU configurations of the edge server and devices are jointly optimized to minimize the long-term average energy requirement of the systems. The key idea is to design a DQN for learning the channel allocation and task admission, while the WPT, offloading time and CPU configurations are efficiently optimized to precisely evaluate the reward of the DQN and substantially reduce its action space. Another important aspect is that a new action generation method is developed to expand and diversify the actions of the DQN, further accelerating its convergence. As validated by simulations, the proposed TS-DQN is much more energy efficient and converges much faster, than its potential alternative directly using the state-of-the-art Deep Deterministic Policy Gradient algorithm to learn all decision variables.
KW - convex optimization
KW - deep Q-network
KW - energy-efficient
KW - Mobile edge computing
KW - resource allocation
KW - wireless power transfer
UR - http://www.scopus.com/inward/record.url?scp=85149376617&partnerID=8YFLogxK
U2 - 10.1109/TCOMM.2023.3251353
DO - 10.1109/TCOMM.2023.3251353
M3 - Article
AN - SCOPUS:85149376617
SN - 0090-6778
VL - 71
SP - 2698
EP - 2710
JO - IEEE Transactions on Communications
JF - IEEE Transactions on Communications
IS - 5
ER -