TY - GEN
T1 - NFVdeep
T2 - 2019 International Symposium on Quality of Service, IWQoS 2019
AU - Xiao, Yikai
AU - Zhang, Qixia
AU - Liu, Fangming
AU - Wang, Jia
AU - Zhao, Miao
AU - Zhang, Zhongxing
AU - Zhang, Jiaxing
N1 - Publisher Copyright:
© 2019 Association for Computing Machinery.
PY - 2019/6/24
Y1 - 2019/6/24
N2 - With the evolution of network function virtualization (NFV), diverse network services can be ?exibly o?ered as service function chains (SFCs) consisted of di?erent virtual network functions (VNFs). However, network state and tra?c typically exhibit unpredictable variations due to stochastically arriving requests with di?erent quality of service (QoS) requirements. Thus, an adaptive online SFC deployment approach is needed to handle the real-time network variations and various service requests. In this paper, we ?rstly introduce a Markov decision process (MDP) model to capture the dynamic network state transitions. In order to jointly minimize the operation cost of NFV providers and maximize the total throughput of requests, we propose NFVdeep, an adaptive, online, deep reinforcement learning approach to automatically deploy SFCs for requests with di?erent QoS requirements. Speci?cally, we use a serialization-and-backtracking method to e?ectively deal with large discrete action space. We also adopt a policy gradient based method to improve the training e?ciency and convergence to optimality. Extensive experimental results demonstrate that NFVdeep converges fast in the training process and responds rapidly to arriving requests especially in large, frequently transferred network state space. Consequently, NFVdeep surpasses the state-of-the-art methods by 32.59% higher accepted throughput and 33.29% lower operation cost on average.
AB - With the evolution of network function virtualization (NFV), diverse network services can be ?exibly o?ered as service function chains (SFCs) consisted of di?erent virtual network functions (VNFs). However, network state and tra?c typically exhibit unpredictable variations due to stochastically arriving requests with di?erent quality of service (QoS) requirements. Thus, an adaptive online SFC deployment approach is needed to handle the real-time network variations and various service requests. In this paper, we ?rstly introduce a Markov decision process (MDP) model to capture the dynamic network state transitions. In order to jointly minimize the operation cost of NFV providers and maximize the total throughput of requests, we propose NFVdeep, an adaptive, online, deep reinforcement learning approach to automatically deploy SFCs for requests with di?erent QoS requirements. Speci?cally, we use a serialization-and-backtracking method to e?ectively deal with large discrete action space. We also adopt a policy gradient based method to improve the training e?ciency and convergence to optimality. Extensive experimental results demonstrate that NFVdeep converges fast in the training process and responds rapidly to arriving requests especially in large, frequently transferred network state space. Consequently, NFVdeep surpasses the state-of-the-art methods by 32.59% higher accepted throughput and 33.29% lower operation cost on average.
KW - Deep Reinforcement Learning
KW - Network Function Virtualization (NFV)
KW - QoS-Aware Resource Management
KW - Service Function Chain
UR - http://www.scopus.com/inward/record.url?scp=85069164250&partnerID=8YFLogxK
U2 - 10.1145/3326285.3329056
DO - 10.1145/3326285.3329056
M3 - Conference Proceeding
AN - SCOPUS:85069164250
T3 - Proceedings of the International Symposium on Quality of Service, IWQoS 2019
BT - Proceedings of the International Symposium on Quality of Service, IWQoS 2019
PB - Association for Computing Machinery, Inc
Y2 - 24 June 2019 through 25 June 2019
ER -