Abstract
Markov decision process (MDP) is commonly used to model a stochastic environment for supporting optimal decision making. However, solving a large-scale MDP problem under the partially observable condition (also called POMDP) is known to be computationally intractable. Belief compression by reducing belief state dimension has recently been shown to be an effective way for making the problem tractable. With the conjecture that temporally close belief states should possess a low intrinsic degree of freedom due to problem regularity, this paper proposes to cluster the belief states based on a criterion function measuring the belief states spatial and temporal differences. Further reduction of the belief state dimension can then result in a more efficient POMDP solver. The proposed method has been tested using a synthesized navigation problem (Hallway2) and empirically shown to be a promising direction towards solving large-scale POMDP problems. Some future research directions are also included.
Original language | English |
---|---|
Pages | 17-24 |
Number of pages | 8 |
Publication status | Published - 2005 |
Event | 5th Workshop on Reasoning with Uncertainty in Robotics, RUR 2005, Held at the International Joint Conference on Artificial Intelligence, IJCAI 2005 - Edinburgh, Scotland, United Kingdom Duration: 30 Jul 2005 → 30 Jul 2005 |
Workshop
Workshop | 5th Workshop on Reasoning with Uncertainty in Robotics, RUR 2005, Held at the International Joint Conference on Artificial Intelligence, IJCAI 2005 |
---|---|
Country/Territory | United Kingdom |
City | Edinburgh, Scotland |
Period | 30/07/05 → 30/07/05 |