Mobile edge computing (MEC) is considered a novel paradigm for computation-intensive and delay-sensitive tasks in fifth generation (5G) networks and beyond. However, its uncertainty, referred to as dynamic and randomness, from the mobile device, wireless channel, and edge network sides, results in high-dimensional, nonconvex, nonlinear, and NP-hard optimization problems. Thanks to the evolved reinforcement learning (RL), upon iteratively interacting with the dynamic and random environment, its trained agent can intelligently obtain the optimal policy in MEC. Furthermore, its evolved versions, such as deep RL (DRL), can achieve higher convergence speed efficiency and learning accuracy based on the parametric approximation for the large-scale state-action space. This paper provides a comprehensive research review on RL-enabled MEC and offers insight for development in this area. More importantly, associated with free mobility, dynamic channels, and distributed services, the MEC challenges that can be solved by different kinds of RL algorithms are identified, followed by how they can be solved by RL solutions in diverse mobile applications. Finally, the open challenges are discussed to provide helpful guidance for future research in RL training and learning MEC.
翻译:移动边缘计算(MEC)被认为是第五代(5G)网络内外计算密集型和延迟敏感任务的新模式,然而,其不确定性,即移动设备、无线频道和边缘网络侧面的动态和随机性,导致高维、非康维克斯、非线性和NP硬优化问题。由于不断发展的强化学习(RL),在与动态和随机环境互动后,其训练有素的代理商可以明智地获得MEC的最佳政策。此外,其演变版本,如深RL(DRL)等,可以在大规模状态行动空间的参数近似基础上,实现更高的趋同速度效率和学习精度。本文对借助RL的MEC提供了全面的研究审查,为这一领域的发展提供了深入了解。更重要的是,与自由流动性、动态渠道和分布式服务相关联,确定了可以通过不同类型RL算法解决的MEC挑战,随后又确定了如何通过不同移动应用的RL解决方案加以解决。最后,讨论了公开的挑战,以便为今后在REC的培训和学习中的研究提供有益的指导。