Structure-aware reinforcement learning for node-overload protection in mobile edge computing

A Jitani, A Mahajan, Z Zhu, H Abou-Zeid… - IEEE Transactions …, 2022 - ieeexplore.ieee.org
IEEE Transactions on Cognitive Communications and Networking, 2022ieeexplore.ieee.org
Mobile Edge Computing (MEC) involves placing computational capability and applications
at the edge of the network, providing benefits such as reduced latency, reduced network
congestion, and improved performance of applications. The performance and reliability of
MEC degrades significantly when the edge server (s) in the cluster are overloaded. In this
work, an adaptive admission control policy to prevent edge node from getting overloaded is
presented. This approach is based on a recently-proposed low complexity RL …
Mobile Edge Computing (MEC) involves placing computational capability and applications at the edge of the network, providing benefits such as reduced latency, reduced network congestion, and improved performance of applications. The performance and reliability of MEC degrades significantly when the edge server(s) in the cluster are overloaded. In this work, an adaptive admission control policy to prevent edge node from getting overloaded is presented. This approach is based on a recently-proposed low complexity RL (Reinforcement Learning) algorithm called SALMUT (Structure-Aware Learning for Multiple Thresholds), which exploits the structure of the optimal admission control policy in multi-class queues for an average-cost setting. We extend the framework to work for node overload-protection problem in a discounted-cost setting. The proposed solution is validated using several scenarios mimicking real-world deployments in two different settings — computer simulations and a docker testbed. Our empirical evaluations show that the total discounted cost incurred by SALMUT is similar to state-of-the-art deep RL algorithms such as PPO (Proximal Policy Optimization) and A2C (Advantage Actor Critic) but requires an order of magnitude less time to train, outputs easily interpretable policy, and can be deployed in an online manner.
ieeexplore.ieee.org