A Deep-Reinforcement-Learning-Based Computation Offloading with Mobile Vehicles in Vehicular Edge Computing

Jie Lin, Siqi Huang, Hanlin Zhang, Xinyu Yang*, Peng Zhao

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

7 Citations (Scopus)

Abstract

Vehicular edge networks involve edge servers that are close to mobile devices to provide extra computation resource to complete the computation tasks of mobile devices with low latency and high reliability. Considerable efforts on computation offloading in vehicular edge networks have been developed to reduce the energy consumption and computation latency, in which roadside units (RSUs) are usually considered as the fixed edge servers (FESs). Nonetheless, the computation offloading with considering mobile vehicles as mobile edge servers (MESs) in vehicular edge networks still needs to be further investigated. To this end, in this article, we propose a Deep-Reinforcement-Learning-based computation offloading with mobile vehicles in vehicular edge computing, namely, Deep-Reinforcement-Learning-based computation offloading scheme (DRL-COMV), in which some vehicles (such as autonomous vehicle) are deployed and considered as the MESs that move in vehicular edge networks and cooperate with FESs to provide extra computation resource for mobile devices, in order to assist in completing the computation tasks of these mobile devices with great Quality of Experience (QoE) (i.e., low latency) for mobile devices. Particularly, the computation offloading model with considering both mobile and FESs is conducted to achieve the computation tasks offloading through vehicle-to-vehicle (V2V) communications, and a collaborative route planning is considered for these MESs to move in vehicular edge networks with objective of improving efficiency of computation offloading. Then, a Deep-Reinforcement-Learning approach with designing rational reward function is proposed to determine the effective computation offloading strategies for multiple mobile devices and multiple edge servers with objective of maximizing both QoE (i.e., low latency) for mobile devices. Through performance evaluations, our results show that our proposed DRL-COMV scheme can achieve a great convergence and stability. Additionally, our results also demonstrate that our DRL-COMV scheme also can achieve better both QoE and task offloading requests hit ratio for mobile devices in comparison with existing approaches (i.e., DDPG, IMOPSOQ, and GABDOS).

Original languageEnglish
Pages (from-to)15501-15514
Number of pages14
JournalIEEE Internet of Things Journal
Volume10
Issue number17
DOIs
Publication statusPublished - 1 Sept 2023
Externally publishedYes

Keywords

  • computation offloading
  • Computational modeling
  • deep reinforcement learning
  • Iron
  • mobile edge servers
  • Mobile handsets
  • Quality of experience
  • Reinforcement learning
  • Servers
  • Task analysis
  • V2V communications
  • Vehicular edge networks
  • Deep-Reinforcement-Learning
  • vehicle-to-vehicle (V2V) communications
  • mobile edge servers (MESs)
  • Computation offloading
  • vehicular edge networks

Fingerprint

Dive into the research topics of 'A Deep-Reinforcement-Learning-Based Computation Offloading with Mobile Vehicles in Vehicular Edge Computing'. Together they form a unique fingerprint.

Cite this