跳到主要导航 跳到搜索 跳到主要内容

Mix-up Consistent Cross Representations for Data-Efficient Reinforcement Learning

  • Shiyu Liu
  • , Guitao Cao*
  • , Yong Liu
  • , Yan Li
  • , Chunwei Wu
  • , Xidong Xi
  • *此作品的通讯作者

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

Deep reinforcement learning (RL) has achieved re-markable performance in sequential decision-making problems. However, it is a challenge for deep RL methods to extract task-relevant semantic information when interacting with limited data from the environment. In this paper, we propose Mix-up Consistent Cross Representations (MCCR), a novel self-supervised auxiliary task, which aims to improve data efficiency and encourage representation prediction. Specifically, we calculate the contrastive loss between low-dimensional and high-dimensional representations of different state observations to boost the mutual information between states, thus improving data efficiency. Furthermore, we employ a mixed strategy to generate intermediate samples, increasing data diversity and the smoothness of representations prediction in nearby timesteps. Experimental results show that MCCR achieves competitive results over the state-of-the-art approaches for complex control tasks in DeepMind Control Suite, notably improving the ability of pretrained encoders to generalize to unseen tasks.

源语言英语
主期刊名2022 International Joint Conference on Neural Networks, IJCNN 2022 - Proceedings
出版商Institute of Electrical and Electronics Engineers Inc.
ISBN(电子版)9781728186719
DOI
出版状态已出版 - 2022
活动2022 International Joint Conference on Neural Networks, IJCNN 2022 - Padua, 意大利
期限: 18 7月 202223 7月 2022

出版系列

姓名Proceedings of the International Joint Conference on Neural Networks
ISSN(印刷版)2161-4393
ISSN(电子版)2161-4407

会议

会议2022 International Joint Conference on Neural Networks, IJCNN 2022
国家/地区意大利
Padua
时期18/07/2223/07/22

指纹

探究 'Mix-up Consistent Cross Representations for Data-Efficient Reinforcement Learning' 的科研主题。它们共同构成独一无二的指纹。

引用此