Attentive multi-view reinforcement learning

  • Yueyue Hu
  • , Shiliang Sun*
  • , Xin Xu
  • , Jing Zhao
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

9 Scopus citations

Abstract

The reinforcement learning process usually takes millions of steps from scratch, due to the limited observation experience. More precisely, the representation approximated by a single deep network is usually limited for reinforcement learning agents. In this paper, we propose a novel multi-view deep attention network (MvDAN), which introduces multi-view representation learning into the reinforcement learning framework for the first time. Based on the multi-view scheme of function approximation, the proposed model approximates multiple view-specific policy or value functions in parallel by estimating the middle-level representation and integrates these functions based on attention mechanisms to generate a comprehensive strategy. Furthermore, we develop the multi-view generalized policy improvement to jointly optimize all policies instead of a single one. Compared with the single-view function approximation scheme in reinforcement learning methods, experimental results on eight Atari benchmarks show that MvDAN outperforms the state-of-the-art methods and has faster convergence and training stability.

Original languageEnglish
Pages (from-to)2461-2474
Number of pages14
JournalInternational Journal of Machine Learning and Cybernetics
Volume11
Issue number11
DOIs
StatePublished - 1 Nov 2020

Keywords

  • Deep reinforcement learning
  • Function approximation
  • Multi-view learning
  • Representation learning

Fingerprint

Dive into the research topics of 'Attentive multi-view reinforcement learning'. Together they form a unique fingerprint.

Cite this