Sciweavers

332 search results - page 14 / 67
» Ranking policies in discrete Markov decision processes
Sort
View
KI
2007
Springer
13 years 7 months ago
Solving Decentralized Continuous Markov Decision Problems with Structured Reward
We present an approximation method that solves a class of Decentralized hybrid Markov Decision Processes (DEC-HMDPs). These DEC-HMDPs have both discrete and continuous state variab...
Emmanuel Benazera
NIPS
2000
13 years 9 months ago
APRICODD: Approximate Policy Construction Using Decision Diagrams
We propose a method of approximate dynamic programming for Markov decision processes (MDPs) using algebraic decision diagrams (ADDs). We produce near-optimal value functions and p...
Robert St-Aubin, Jesse Hoey, Craig Boutilier
NIPS
2008
13 years 9 months ago
Particle Filter-based Policy Gradient in POMDPs
Our setting is a Partially Observable Markov Decision Process with continuous state, observation and action spaces. Decisions are based on a Particle Filter for estimating the bel...
Pierre-Arnaud Coquelin, Romain Deguest, Rém...
VTC
2008
IEEE
173views Communications» more  VTC 2008»
14 years 2 months ago
Adaptive Call Admission Control with Dynamic Resource Reallocation for Cell-Based Multirate Wireless Systems
—This paper studies the admission control and resource allocation in a cell-based wireless system that supports singlemedia and multirate services. Utilizing the idea of adaptive...
Kai-Wei Ke, Chen-Nien Tsai, Ho-Ting Wu, Chia-Hao H...
GECCO
2009
Springer
162views Optimization» more  GECCO 2009»
13 years 5 months ago
Uncertainty handling CMA-ES for reinforcement learning
The covariance matrix adaptation evolution strategy (CMAES) has proven to be a powerful method for reinforcement learning (RL). Recently, the CMA-ES has been augmented with an ada...
Verena Heidrich-Meisner, Christian Igel