Sciweavers

74 search results - page 8 / 15
» Regret Bounds for Gaussian Process Bandit Problems
Sort
View
CIMCA
2008
IEEE
14 years 2 months ago
Tree Exploration for Bayesian RL Exploration
Research in reinforcement learning has produced algorithms for optimal decision making under uncertainty that fall within two main types. The first employs a Bayesian framework, ...
Christos Dimitrakakis
NIPS
2004
13 years 9 months ago
Experts in a Markov Decision Process
We consider an MDP setting in which the reward function is allowed to change during each time step of play (possibly in an adversarial manner), yet the dynamics remain fixed. Simi...
Eyal Even-Dar, Sham M. Kakade, Yishay Mansour
CORR
2010
Springer
105views Education» more  CORR 2010»
13 years 6 months ago
Optimism in Reinforcement Learning Based on Kullback-Leibler Divergence
We consider model-based reinforcement learning in finite Markov Decision Processes (MDPs), focussing on so-called optimistic strategies. Optimism is usually implemented by carryin...
Sarah Filippi, Olivier Cappé, Aurelien Gari...
JMLR
2010
155views more  JMLR 2010»
13 years 2 months ago
Bayesian Gaussian Process Latent Variable Model
We introduce a variational inference framework for training the Gaussian process latent variable model and thus performing Bayesian nonlinear dimensionality reduction. This method...
Michalis Titsias, Neil D. Lawrence
JAT
2007
86views more  JAT 2007»
13 years 7 months ago
Gaussian averages of interpolated bodies and applications to approximate reconstruction
We prove sharp bounds for the expectation of the supremum of the Gaussian process indexed by the intersection of Bn p with ρBn q for 1 ≤ p, q ≤ ∞ and ρ > 0, and by the ...
Y. Gordon, A. E. Litvak, Shahar Mendelson, A. Pajo...