Arrow Research search
Back to EWRL

EWRL 2024

Adaptive Exploration for Data-Efficient General Value Function Evaluations

Workshop Paper EWRL17 Artificial Intelligence · Machine Learning · Reinforcement Learning

Abstract

General Value Functions (GVFs) (Sutton et al. , 2011) represent predictive knowledge in reinforcement learning. Each GVF computes the expected return for a given policy, based on a unique reward. Existing methods relying on fixed behavior policies or pre-collected data often face data efficiency issues when learning multiple GVFs in parallel using off-policy methods. To address this, we introduce $GVFExplorer$ which adaptively learns a single behavior policy that efficiently collects data for evaluating multiple GVFs in parallel. Our method optimizes the behavior policy by minimizing the total variance in return across GVFs, thereby reducing the required environmental interactions We use an existing temporal-difference-style variance estimator to approximate the return variance. We prove that each behavior policy update decreases the overall mean squared error in GVF predictions. We empirically show our method's performance in tabular and nonlinear function approximation settings, including Mujoco environments, with stationary and non-stationary reward signals, optimizing data usage and reducing prediction errors across multiple GVFs.

Authors

Keywords

  • exploration for GVFs
  • general value functions
  • GVFs
  • multiple policy evaluations
  • variance-minimization

Context

Venue
European Workshop on Reinforcement Learning
Archive span
2008-2025
Indexed papers
649
Paper id
663020702526214600