Improving value function approximation in factored POMDPs by exploiting model structure

Tiago S. Veiga, Matthijs T.J. Spaan, Pedro U. Lima

Research output: Chapter in Book/Conference proceedings/Edited volumeConference contributionScientificpeer-review

Abstract

Linear value function approximation in Markov decision processes (MDPs) has been studied extensively, but there are several challenges when applying such techniques to partially observable MDPs (POMDPs). Furthermore, the system designer often has to choose a set of basis functions. We propose an automatic method to derive a suitable set of basis functions by exploiting the structure of factored models. We experimentally show that our approximation can reduce the solution size by several orders of magnitude in large problems.

Original languageEnglish
Title of host publicationAAMAS 2015 - Proceedings of the 2015 International Conference on Autonomous Agents and Multiagent Systems
PublisherInternational Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS)
Pages1827-1828
Number of pages2
Volume3
ISBN (Electronic)9781450337717
Publication statusPublished - 1 Jan 2015
EventAAMAS 2015: 14th International Conference on Autonomous Agents and Multiagent Systems - Istanbul, Turkey
Duration: 4 May 20158 May 2015
Conference number: 14

Conference

ConferenceAAMAS 2015: 14th International Conference on Autonomous Agents and Multiagent Systems
Abbreviated titleAAMAS 2015
Country/TerritoryTurkey
CityIstanbul
Period4/05/158/05/15

Keywords

  • POMDP
  • Value function approximation

Fingerprint

Dive into the research topics of 'Improving value function approximation in factored POMDPs by exploiting model structure'. Together they form a unique fingerprint.

Cite this