Reinforcement Learning-Empowered Mobile Edge Computing for 6G Edge Intelligence

  • Peng Wei
  • , Kun Guo
  • , Ye Li
  • , Jue Wang
  • , Wei Feng*
  • , Shi Jin
  • , Ning Ge
  • , Ying Chang Liang
  • *Corresponding author for this work

Research output: Contribution to journalReview articlepeer-review

57 Scopus citations

Abstract

Mobile edge computing (MEC) is considered a novel paradigm for computation-intensive and delay-sensitive tasks in fifth generation (5G) networks and beyond. However, its uncertainty, referred to as dynamic and randomness, from the mobile device, wireless channel, and edge network sides, results in high-dimensional, nonconvex, nonlinear, and NP-hard optimization problems. Thanks to the evolved reinforcement learning (RL), upon iteratively interacting with the dynamic and random environment, its trained agent can intelligently obtain the optimal policy in MEC. Furthermore, its evolved versions, such as deep reinforcement learning (DRL), can achieve higher convergence speed efficiency and learning accuracy based on the parametric approximation for the large-scale state-action space. This paper provides a comprehensive research review on RL-enabled MEC and offers insight for development in this area. More importantly, associated with free mobility, dynamic channels, and distributed services, the MEC challenges that can be solved by different kinds of RL algorithms are identified, followed by how they can be solved by RL solutions in diverse mobile applications. Finally, the open challenges are discussed to provide helpful guidance for future research in RL training and learning MEC.

Original languageEnglish
Pages (from-to)65156-65192
Number of pages37
JournalIEEE Access
Volume10
DOIs
StatePublished - 2022

Keywords

  • Mobile edge computing (MEC)
  • network uncertainty
  • reinforcement learning (RL)

Fingerprint

Dive into the research topics of 'Reinforcement Learning-Empowered Mobile Edge Computing for 6G Edge Intelligence'. Together they form a unique fingerprint.

Cite this