TY - CHAP
T1 - Dynamic VNF Resource Scaling and Migration
T2 - A Machine Learning Approach
AU - Zhuang, Weihua
AU - Qu, Kaige
N1 - Publisher Copyright:
© 2021, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2021
Y1 - 2021
N2 - In this chapter, VNF scalability via joint resource scaling and migration is studied in a local network segment, to meet a probabilistic delay requirement in the presence of non-stationary traffic with changing traffic statistics. A change point detection scheme determines boundaries between stationary traffic segments in an online manner as new traffic samples arrive, and provides a triggering signal for VNF scalability. Under the fBm traffic model assumption, the resource demand for a probabilistic delay guarantee is predicted for each newly detected stationary traffic segment, based on traffic parameter learning with Gaussian process regression and fBm resource provisioning model. Given the predicted resource demand, a dynamic VNF migration problem is formulated as a Markov decision process (MDP) with variable-length decision epochs, to maximize the long-term reward integrating load balancing, migration cost, and resource overloading penalty. The MDP is solved by a reinforcement leaning (RL) approach. Specifically, a penalty-aware deep Q -learning algorithm demonstrating advantages in reducing training loss and increasing cumulative reward is employed to learn the adaptive VNF migration actions under the dynamics in change points, resource demand, and background traffic.
AB - In this chapter, VNF scalability via joint resource scaling and migration is studied in a local network segment, to meet a probabilistic delay requirement in the presence of non-stationary traffic with changing traffic statistics. A change point detection scheme determines boundaries between stationary traffic segments in an online manner as new traffic samples arrive, and provides a triggering signal for VNF scalability. Under the fBm traffic model assumption, the resource demand for a probabilistic delay guarantee is predicted for each newly detected stationary traffic segment, based on traffic parameter learning with Gaussian process regression and fBm resource provisioning model. Given the predicted resource demand, a dynamic VNF migration problem is formulated as a Markov decision process (MDP) with variable-length decision epochs, to maximize the long-term reward integrating load balancing, migration cost, and resource overloading penalty. The MDP is solved by a reinforcement leaning (RL) approach. Specifically, a penalty-aware deep Q -learning algorithm demonstrating advantages in reducing training loss and increasing cumulative reward is employed to learn the adaptive VNF migration actions under the dynamics in change points, resource demand, and background traffic.
UR - https://www.scopus.com/pages/publications/85118803794
U2 - 10.1007/978-3-030-87136-9_4
DO - 10.1007/978-3-030-87136-9_4
M3 - 章节
AN - SCOPUS:85118803794
T3 - Wireless Networks (United Kingdom)
SP - 85
EP - 129
BT - Wireless Networks (United Kingdom)
PB - Springer Science and Business Media B.V.
ER -