TY - JOUR
T1 - Wireless Resource Scheduling in Virtualized Radio Access Networks Using Stochastic Learning
AU - Chen, Xianfu
AU - Han, Zhu
AU - Zhang, Honggang
AU - Xue, Guoliang
AU - Xiao, Yong
AU - Bennis, Mehdi
N1 - Funding Information:
This research was supported in part by AKA grants 310786 and 289611, TEKES grants 2364/31/2014 and 2368/31/2014, US National Science Foundation grants 1717454, 1731424, 1704092, 1702850, 1646607, 1547201, 1434789, 1456921, 1443917, 1405121, 1457262 and 1461886, and the Program for Zhejiang Leading Team of Science and Technology Innovation under Grant 2013TD20.
Publisher Copyright:
© 2002-2012 IEEE.
PY - 2018/4/1
Y1 - 2018/4/1
N2 - How to allocate the limited wireless resource in dense radio access networks (RANs) remains challenging. By leveraging a software-defined control plane, the independent base stations (BSs) are virtualized as a centralized network controller (CNC). Such virtualization decouples the CNC from the wireless service providers (WSPs). We investigate a virtualized RAN, where the CNC auctions channels at the beginning of scheduling slots to the mobile terminals (MTs) based on bids from their subscribing WSPs. Each WSP aims at maximizing the expected long-term payoff from bidding channels to satisfy the MTs for transmitting packets. We formulate the problem as a stochastic game, where the channel auction and packet scheduling decisions of a WSP depend on the state of network and the control policies of its competitors. To approach the equilibrium solution, an abstract stochastic game is proposed with bounded regret. The decision making process of each WSP is modeled as a Markov decision process (MDP). To address the signalling overhead and computational complexity issues, we decompose the MDP into a series of single-agent MDPs with reduced state spaces, and derive an online localized algorithm to learn the state value functions. Our results show significant performance improvements in terms of per-MT average utility.
AB - How to allocate the limited wireless resource in dense radio access networks (RANs) remains challenging. By leveraging a software-defined control plane, the independent base stations (BSs) are virtualized as a centralized network controller (CNC). Such virtualization decouples the CNC from the wireless service providers (WSPs). We investigate a virtualized RAN, where the CNC auctions channels at the beginning of scheduling slots to the mobile terminals (MTs) based on bids from their subscribing WSPs. Each WSP aims at maximizing the expected long-term payoff from bidding channels to satisfy the MTs for transmitting packets. We formulate the problem as a stochastic game, where the channel auction and packet scheduling decisions of a WSP depend on the state of network and the control policies of its competitors. To approach the equilibrium solution, an abstract stochastic game is proposed with bounded regret. The decision making process of each WSP is modeled as a Markov decision process (MDP). To address the signalling overhead and computational complexity issues, we decompose the MDP into a series of single-agent MDPs with reduced state spaces, and derive an online localized algorithm to learn the state value functions. Our results show significant performance improvements in terms of per-MT average utility.
KW - Learning
KW - Markov decision process
KW - Multi-user resource scheduling
KW - Network virtualization
KW - Radio access networks
KW - Software-defined networking
KW - Stochastic games
UR - http://www.scopus.com/inward/record.url?scp=85028500008&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85028500008&partnerID=8YFLogxK
U2 - 10.1109/TMC.2017.2742949
DO - 10.1109/TMC.2017.2742949
M3 - Article
AN - SCOPUS:85028500008
SN - 1536-1233
VL - 17
SP - 961
EP - 974
JO - IEEE Transactions on Mobile Computing
JF - IEEE Transactions on Mobile Computing
IS - 4
ER -