@inproceedings{49fd73e6f61c47c1ab4a894d8691197d,
title = "HMRL: Hyper-Meta Learning for Sparse Reward Reinforcement Learning Problem",
abstract = "In spite of the success of existing meta reinforcement learning methods, they still have difficulty in learning a meta policy effectively for RL problems with sparse reward. In this respect, we develop a novel meta reinforcement learning framework called Hyper-Meta RL(HMRL), for sparse reward RL problems. It is consisted with three modules including the cross-environment meta state embedding module which constructs a common meta state space to adapt to different environments; the meta state based environment-specific meta reward shaping which effectively extends the original sparse reward trajectory by cross-environmental knowledge complementarity and as a consequence the meta policy achieves better generalization and efficiency with the shaped meta reward. Experiments with sparse-reward environments show the superiority of HMRL on both transferability and policy learning efficiency.",
keywords = "meta learning, reinforcement learning, sparse reward",
author = "Yun Hua and Xiangfeng Wang and Bo Jin and Wenhao Li and Junchi Yan and Xiaofeng He and Hongyuan Zha",
note = "Publisher Copyright: {\textcopyright} 2021 ACM.; 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD 2021 ; Conference date: 14-08-2021 Through 18-08-2021",
year = "2021",
month = aug,
day = "14",
doi = "10.1145/3447548.3467242",
language = "英语",
series = "Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining",
publisher = "Association for Computing Machinery",
pages = "637--645",
booktitle = "KDD 2021 - Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining",
address = "美国",
}