A Closer Look at MOMDPs
Abstract
The difficulties encountered in sequential decision-making problems under uncertainty are often linked to the large size of the state space. Exploiting the structure of the problem, for example by employing a factored representation, is usually an efficient approach but, in the case of partially observable Markov decision processes, the fact that some state variables may be visible has not been sufficiently appreciated. In this article, we present a complementary analysis and discussion about MOMDPs, a formalism that exploits the fact that the state space may be factored in one visible part and one hidden part. Starting from a POMDP description, we dig into the structure of the belief update, value function, and the consequences in value iteration, specifically how classical algorithms can be adapted to this factorization, and demonstrate the resulting benefits through an empirical evaluation.
Más información
Título según WOS: | ID WOS:000287040000028 Not found in local WOS DB |
Título de la Revista: | 22ND INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2010), PROCEEDINGS, VOL 2 |
Editorial: | IEEE |
Fecha de publicación: | 2010 |
Página de inicio: | 197 |
Página final: | 204 |
DOI: |
10.1109/ICTAI.2010.101 |
Notas: | ISI |