Sciweavers

260 search results - page 41 / 52
» Quasi-Deterministic Partially Observable Markov Decision Pro...
Sort
View
AIED
2011
Springer
13 years 12 days ago
Faster Teaching by POMDP Planning
Both human and automated tutors must infer what a student knows and plan future actions to maximize learning. Though substantial research has been done on tracking and modeling stu...
Anna N. Rafferty, Emma Brunskill, Thomas L. Griffi...
AAAI
2006
13 years 10 months ago
Compact, Convex Upper Bound Iteration for Approximate POMDP Planning
Partially observable Markov decision processes (POMDPs) are an intuitive and general way to model sequential decision making problems under uncertainty. Unfortunately, even approx...
Tao Wang, Pascal Poupart, Michael H. Bowling, Dale...
CISS
2008
IEEE
14 years 3 months ago
Rate adaptation via link-layer feedback for goodput maximization over a time-varying channel
Abstract—We consider adapting the transmission rate to maximize the goodput, i.e., the amount of data transmitted without error, over a continuous Markov flat-fading wireless ch...
Rohit Aggarwal, Phil Schniter, Can Emre Koksal
AI
2006
Springer
14 years 16 days ago
Belief Selection in Point-Based Planning Algorithms for POMDPs
Abstract. Current point-based planning algorithms for solving partially observable Markov decision processes (POMDPs) have demonstrated that a good approximation of the value funct...
Masoumeh T. Izadi, Doina Precup, Danielle Azar
GECCO
2009
Springer
162views Optimization» more  GECCO 2009»
13 years 6 months ago
Uncertainty handling CMA-ES for reinforcement learning
The covariance matrix adaptation evolution strategy (CMAES) has proven to be a powerful method for reinforcement learning (RL). Recently, the CMA-ES has been augmented with an ada...
Verena Heidrich-Meisner, Christian Igel