Arrow Research search
Back to AAMAS

AAMAS 2025

Tighter Value-Function Approximations for POMDPs

Conference Paper Research Paper Track Autonomous Agents and Multiagent Systems

Abstract

Solving partially observable Markov decision processes (POMDPs) typically requires reasoning about the values of exponentially many state beliefs. Towards practical performance, state-of-the-art solvers use value bounds to guide this reasoning. However, sound upper value bounds are often computationally expensive to compute, and there is a tradeoff between the tightness of such bounds and their computational cost. This paper introduces new and provably tighter upper value bounds than the commonly used fast informed bound. Our empirical evaluation shows that, despite their additional computational overhead, the new upper bounds accelerate state-ofthe-art POMDP solvers on a wide range of benchmarks.

Authors

Keywords

  • POMDPs
  • Heuristic Search
  • Value Bounds
  • Planning

Context

Venue
International Conference on Autonomous Agents and Multiagent Systems
Archive span
2002-2025
Indexed papers
7403
Paper id
159642532052009730