Sciweavers

67 search results - page 8 / 14
» Limits of Multi-Discounted Markov Decision Processes
Sort
View
ICC
2008
IEEE
169views Communications» more  ICC 2008»
14 years 2 months ago
Optimality of Myopic Sensing in Multi-Channel Opportunistic Access
—We consider opportunistic communications over multiple channels where the state (“good” or “bad”) of each channel evolves as independent and identically distributed Mark...
Tara Javidi, Bhaskar Krishnamachari, Qing Zhao, Mi...
IJCAI
2003
13 years 9 months ago
Automated Generation of Understandable Contingency Plans
Markov decision processes (MDPs) and contingency planning (CP) are two widely used approaches to planning under uncertainty. MDPs are attractive because the model is extremely gen...
Max Horstmann, Shlomo Zilberstein
ATAL
2008
Springer
13 years 9 months ago
Not all agents are equal: scaling up distributed POMDPs for agent networks
Many applications of networks of agents, including mobile sensor networks, unmanned air vehicles, autonomous underwater vehicles, involve 100s of agents acting collaboratively und...
Janusz Marecki, Tapana Gupta, Pradeep Varakantham,...
NIPS
2003
13 years 9 months ago
Approximate Policy Iteration with a Policy Language Bias
We study an approach to policy selection for large relational Markov Decision Processes (MDPs). We consider a variant of approximate policy iteration (API) that replaces the usual...
Alan Fern, Sung Wook Yoon, Robert Givan
ICML
1995
IEEE
14 years 8 months ago
Learning Policies for Partially Observable Environments: Scaling Up
Partially observable Markov decision processes (pomdp's) model decision problems in which an agent tries to maximize its reward in the face of limited and/or noisy sensor fee...
Michael L. Littman, Anthony R. Cassandra, Leslie P...