Autonomous Resource Slicing for Virtualized Vehicular Networks with D2D Communications Based on Deep Reinforcement Learning

Guolin Sun*, Gordon Owusu Boateng, Daniel Ayepah-Mensah, Guisong Liu, Jiang Wei

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

50 Citations (Scopus)

Abstract

Considering bandwidth-hungry and low latency requirement of vehicular communications applications, we propose a novel dynamic reinforcement learning-based slicing framework and optimization solutions for efficient resource provisioning in virtualized network for D2D-based vehicle-to-vehicle (V2V) communication. The aim is to balance resource utilization and quality of service (QoS) satisfaction levels for multiple slices. The slicing framework is designed as a three-stage layered framework. In the first stage, we propose dynamic deep reinforcement learning-based virtual resource allocation scheme to allocate distinct resources to slices. In the second stage, we aggregate the D2D resource portion of the slice resource for D2D-based V2V communication. In the third stage, due to the computational complexity and signaling overhead of the physical resource allocation, we transform the problem into a convex optimization problem and solve with an alternating direction method of multipliers-based distributed algorithm. Performance results are provided in terms of resource utilization, QoS satisfaction and throughput to show the benefit of integrating resource slices dedicated to supporting interslice D2D-based V2V communication in vehicular network.

Original languageEnglish
Article number9070169
Pages (from-to)4694-4705
Number of pages12
JournalIEEE Systems Journal
Volume14
Issue number4
DOIs
Publication statusPublished - Dec 2020
Externally publishedYes

Keywords

  • Deep reinforcement learning (DRL)
  • network slicing
  • resource aggregation
  • resource allocation
  • V2V communication

Fingerprint

Dive into the research topics of 'Autonomous Resource Slicing for Virtualized Vehicular Networks with D2D Communications Based on Deep Reinforcement Learning'. Together they form a unique fingerprint.

Cite this