Arrow Research search
Back to ICLR

ICLR 2021

Representation Balancing Offline Model-based Reinforcement Learning

Conference Paper Poster Presentations Artificial Intelligence ยท Machine Learning

Abstract

One of the main challenges in offline and off-policy reinforcement learning is to cope with the distribution shift that arises from the mismatch between the target policy and the data collection policy. In this paper, we focus on a model-based approach, particularly on learning the representation for a robust model of the environment under the distribution shift, which has been first studied by Representation Balancing MDP (RepBM). Although this prior work has shown promising results, there are a number of shortcomings that still hinder its applicability to practical tasks. In particular, we address the curse of horizon exhibited by RepBM, rejecting most of the pre-collected data in long-term tasks. We present a new objective for model learning motivated by recent advances in the estimation of stationary distribution corrections. This effectively overcomes the aforementioned limitation of RepBM, as well as naturally extending to continuous action spaces and stochastic policies. We also present an offline model-based policy optimization using this new objective, yielding the state-of-the-art performance in a representative set of benchmark offline RL tasks.

Authors

Keywords

  • Reinforcement Learning
  • Model-based Reinforcement Learning
  • Offline Reinforcement Learning
  • Batch Reinforcement Learning
  • Off-policy policy evaluation

Context

Venue
International Conference on Learning Representations
Archive span
2013-2025
Indexed papers
10294
Paper id
877174792227520527