TY - JOUR
T1 - Adaptive Edge Task Offloading via Parameterized Multi-Objective Reinforcement Learning With Hybrid Action Space
AU - Tong, Huimin
AU - Chen, Cheng
AU - Jiang, Weihao
AU - Wang, Ting
AU - Zhu, Jiang
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2025
Y1 - 2025
N2 - In 6G networks, Multi-access Edge Computing (MEC) enables ultra-low latency and high reliability for Internet of Things (IoT) applications. However, optimizing resource allocation in MEC is challenging due to dynamic network conditions and limited computational resources. To address these challenges, this study proposes a Hybrid Multi-Objective Soft Actor-Critic (HMO-SAC) algorithm, which integrates Multi-Objective Reinforcement Learning (MORL) within a hybrid action space. The method dynamically balances multiple optimization objectives, leveraging a hybrid action space to make decisions involving both discrete and continuous parameters, such as task offloading targets and resource allocation. Additionally, an Improved Near-on Experience Replay (INER) mechanism is introduced to mitigate extrapolation errors in off-policy sampled data. Simulation results demonstrate that HMO-SAC improves convergence speed by 14% on average and reduces the task completion time and energy consumption by 23% compared to state-of-the-art methods.
AB - In 6G networks, Multi-access Edge Computing (MEC) enables ultra-low latency and high reliability for Internet of Things (IoT) applications. However, optimizing resource allocation in MEC is challenging due to dynamic network conditions and limited computational resources. To address these challenges, this study proposes a Hybrid Multi-Objective Soft Actor-Critic (HMO-SAC) algorithm, which integrates Multi-Objective Reinforcement Learning (MORL) within a hybrid action space. The method dynamically balances multiple optimization objectives, leveraging a hybrid action space to make decisions involving both discrete and continuous parameters, such as task offloading targets and resource allocation. Additionally, an Improved Near-on Experience Replay (INER) mechanism is introduced to mitigate extrapolation errors in off-policy sampled data. Simulation results demonstrate that HMO-SAC improves convergence speed by 14% on average and reduces the task completion time and energy consumption by 23% compared to state-of-the-art methods.
KW - Hybrid action space
KW - multi-access edge computing (MEC)
KW - multi-objective reinforcement learning (MORL)
KW - task offloading
UR - https://www.scopus.com/pages/publications/105008021524
U2 - 10.1109/TNSE.2025.3577628
DO - 10.1109/TNSE.2025.3577628
M3 - 文章
AN - SCOPUS:105008021524
SN - 2327-4697
VL - 12
SP - 4876
EP - 4893
JO - IEEE Transactions on Network Science and Engineering
JF - IEEE Transactions on Network Science and Engineering
IS - 6
ER -