Sciweavers

32 search results - page 1 / 7
» Learning Policies for Partially Observable Environments: Sca...
Sort
View
ICML
1995
IEEE
14 years 8 months ago
Learning Policies for Partially Observable Environments: Scaling Up
Partially observable Markov decision processes (pomdp's) model decision problems in which an agent tries to maximize its reward in the face of limited and/or noisy sensor fee...
Michael L. Littman, Anthony R. Cassandra, Leslie P...
AAAI
2007
13 years 9 months ago
Scaling Up: Solving POMDPs through Value Based Clustering
Partially Observable Markov Decision Processes (POMDPs) provide an appropriately rich model for agents operating under partial knowledge of the environment. Since finding an opti...
Yan Virin, Guy Shani, Solomon Eyal Shimony, Ronen ...
ICMLA
2008
13 years 9 months ago
A Predictive Model for Imitation Learning in Partially Observable Environments
Learning by imitation has shown to be a powerful paradigm for automated learning in autonomous robots. This paper presents a general framework of learning by imitation for stochas...
Abdeslam Boularias
ATAL
2008
Springer
13 years 9 months ago
Not all agents are equal: scaling up distributed POMDPs for agent networks
Many applications of networks of agents, including mobile sensor networks, unmanned air vehicles, autonomous underwater vehicles, involve 100s of agents acting collaboratively und...
Janusz Marecki, Tapana Gupta, Pradeep Varakantham,...
ATAL
2009
Springer
14 years 2 months ago
SarsaLandmark: an algorithm for learning in POMDPs with landmarks
Reinforcement learning algorithms that use eligibility traces, such as Sarsa(λ), have been empirically shown to be effective in learning good estimated-state-based policies in pa...
Michael R. James, Satinder P. Singh