Sciweavers

683 search results - page 70 / 137
» Coarticulation in Markov Decision Processes
Sort
View
AIPS
2007
13 years 11 months ago
Prioritizing Bellman Backups without a Priority Queue
Several researchers have shown that the efficiency of value iteration, a dynamic programming algorithm for Markov decision processes, can be improved by prioritizing the order of...
Peng Dai, Eric A. Hansen
AIPS
2008
13 years 11 months ago
Multiagent Planning Under Uncertainty with Stochastic Communication Delays
We consider the problem of cooperative multiagent planning under uncertainty, formalized as a decentralized partially observable Markov decision process (Dec-POMDP). Unfortunately...
Matthijs T. J. Spaan, Frans A. Oliehoek, Nikos A. ...
ATAL
2008
Springer
13 years 11 months ago
Reinforcement learning for DEC-MDPs with changing action sets and partially ordered dependencies
Decentralized Markov decision processes are frequently used to model cooperative multi-agent systems. In this paper, we identify a subclass of general DEC-MDPs that features regul...
Thomas Gabel, Martin A. Riedmiller
AAAI
2006
13 years 10 months ago
Using Homomorphisms to Transfer Options across Continuous Reinforcement Learning Domains
We examine the problem of Transfer in Reinforcement Learning and present a method to utilize knowledge acquired in one Markov Decision Process (MDP) to bootstrap learning in a mor...
Vishal Soni, Satinder P. Singh
IJCAI
2003
13 years 10 months ago
Taming Decentralized POMDPs: Towards Efficient Policy Computation for Multiagent Settings
The problem of deriving joint policies for a group of agents that maximize some joint reward function can be modeled as a decentralized partially observable Markov decision proces...
Ranjit Nair, Milind Tambe, Makoto Yokoo, David V. ...