TY - GEN
T1 - Adaptive Multi-objective Reinforcement Learning for Pareto Frontier Approximation
T2 - 29th European Signal Processing Conference, EUSIPCO 2021
AU - Chen, Ruiqing
AU - Sun, Fanglei
AU - Chen, Liang
AU - Li, Kai
AU - Wu, Liantao
AU - Wang, Jun
AU - Yang, Yang
N1 - Publisher Copyright:
© 2021 European Signal Processing Conference. All rights reserved.
PY - 2021
Y1 - 2021
N2 - Multi-Objective Optimization (MOO) has always been an important issue in the field of wireless communications. With the development of 5G networks, more objectives have been concerned to improve the user experience. The relationship between these multiple objectives is complex or even conflicting, which increases the difficulty of solving the MOO problems. Traditional multi-objective optimization algorithms (e.g., genetic algorithm) have higher computation complexity and require to store multiple models for the preference of different objectives. Therefore, in this paper, a multi-objective scheduling model based on the Actor-Critic framework is proposed, which can effectively solve the multi-user scheduling problem under Massive Multiple-Input Multiple-Output (MIMO), and utilize a single model to approximate the Pareto frontier. In the single-cell downlink scheduling scenario, the proposed model is applied to the two objective optimization, i.e., channel capacity and fairness. The simulation results show that the performance of our model is close to the theoretical optimal value in the single-objective case. The Pareto frontier can be uniformly approximated in the multi-objective case, and it has strong robustness to never-seen preference combinations.
AB - Multi-Objective Optimization (MOO) has always been an important issue in the field of wireless communications. With the development of 5G networks, more objectives have been concerned to improve the user experience. The relationship between these multiple objectives is complex or even conflicting, which increases the difficulty of solving the MOO problems. Traditional multi-objective optimization algorithms (e.g., genetic algorithm) have higher computation complexity and require to store multiple models for the preference of different objectives. Therefore, in this paper, a multi-objective scheduling model based on the Actor-Critic framework is proposed, which can effectively solve the multi-user scheduling problem under Massive Multiple-Input Multiple-Output (MIMO), and utilize a single model to approximate the Pareto frontier. In the single-cell downlink scheduling scenario, the proposed model is applied to the two objective optimization, i.e., channel capacity and fairness. The simulation results show that the performance of our model is close to the theoretical optimal value in the single-objective case. The Pareto frontier can be uniformly approximated in the multi-objective case, and it has strong robustness to never-seen preference combinations.
KW - Massive MIMO
KW - Multi-objective reinforcement learning (MORL)
KW - Pareto frontier
KW - Single cell Multi-User (MU)MIMO scheduling
UR - https://www.scopus.com/pages/publications/85123177299
U2 - 10.23919/EUSIPCO54536.2021.9615934
DO - 10.23919/EUSIPCO54536.2021.9615934
M3 - 会议稿件
AN - SCOPUS:85123177299
T3 - European Signal Processing Conference
SP - 1631
EP - 1635
BT - 29th European Signal Processing Conference, EUSIPCO 2021 - Proceedings
PB - European Signal Processing Conference, EUSIPCO
Y2 - 23 August 2021 through 27 August 2021
ER -