Solving POMDPs with continuous or large discrete observation spaces

Jesse Hoey, Pascal Poupart

    Research output: Non-textual formDigital or Visual Products

    Abstract

    We describe methods to solve partially observable Markov decision processes (POMDPs) with continuous or large discrete observation spaces. Realistic problems often have rich observation spaces, posing significant problems for standard POMDP algorithms that require explicit enumeration of the observations. This problem is usually approached by imposing an a priori discretisation on the observation space, which can be sub-optimal for the decision making task. However, since only those observations that would change the policy need to be distinguished, the decision problem itself induces a lossless partitioning of the observation space. This paper demonstrates how to find this partition while computing a policy, and how the resulting discretisation of the observation space reveals the relevant features of the application domain. The algorithms are demonstrated on a toy example and on a realistic assisted living task.
    Original languageEnglish
    PublisherMorgan Kaufmann
    Publication statusPublished - 2005

    Fingerprint

    Dive into the research topics of 'Solving POMDPs with continuous or large discrete observation spaces'. Together they form a unique fingerprint.

    Cite this