Arrow Research search
Back to EUMAS

EUMAS 2011

Reinforcement Learning Transfer Using a Sparse Coded Inter-task Mapping

Conference Paper European Workshop on Multi-Agent Systems Artificial Intelligence · Multi-Agent Systems

Abstract

Abstract Reinforcement learning agents can successfully learn in a variety of difficult tasks. A fundamental problem is that they may learn slowly in complex environments, inspiring the development of speedup methods such as transfer learning. Transfer improves learning by reusing learned behaviors in similar tasks, usually via an inter-task mapping, which defines how a pair of tasks are related. This paper proposes a novel transfer learning technique to autonomously construct an inter-task mapping by using a novel combinations of sparse coding, sparse projection learning, and sparse pseudo-input gaussian processes. Experiments show successful transfer of information between two very different domains: the mountain car and the pole swing-up task. This paper empirically shows that the learned inter-task mapping can be used to successfully (1) improve the performance of a learned policy on a fixed number of samples, (2) reduce the learning times needed by the algorithms to converge to a policy on a fixed number of samples, and (3) converge faster to a near-optimal policy given a large amount of samples.

Authors

Keywords

No keywords are indexed for this paper.

Context

Venue
European Conference on Multi-Agent Systems
Archive span
2005-2025
Indexed papers
516
Paper id
739139853170573948