Conference Papers

Improving Value Function Approximation in Factored POMDPs by Exploiting Model Structure (Extended Abstract)

Abstract:
Linear value function approximation in Markov decision processes (MDPs) has been studied extensively, but there are several challenges when applying such techniques to partially observable MDPs (POMDPs). Furthermore, the system designer often has to choose a set of basis functions. We propose an automatic method to derive a suitable set of basis functions by exploiting the structure of factored models. We experimentally show that our approximation can reduce the solution size by several orders of magnitude in large problems.
Impact factor:
NULL
URL:
http://dl.acm.org/citation.cfm?id=2773457

Proc. of AAMAS 2015 - 14th International Conference on Autonomous Agents and Multiagent Systems, Istanbul, Turkey, pp. 1827-1828