Sciweavers

201 search results - page 40 / 41
» Solving Concurrent Markov Decision Processes
Sort
View
FOCS
2003
IEEE
14 years 21 days ago
Approximation Algorithms for Orienteering and Discounted-Reward TSP
In this paper, we give the rst constant-factor approximationalgorithmfor the rooted Orienteering problem, as well as a new problem that we call the Discounted-Reward TSP, motivate...
Avrim Blum, Shuchi Chawla, David R. Karger, Terran...
ICDCS
2010
IEEE
13 years 11 months ago
Stochastic Steepest-Descent Optimization of Multiple-Objective Mobile Sensor Coverage
—We propose a steepest descent method to compute optimal control parameters for balancing between multiple performance objectives in stateless stochastic scheduling, wherein the ...
Chris Y. T. Ma, David K. Y. Yau, Nung Kwan Yip, Na...
IJRR
2011
218views more  IJRR 2011»
13 years 2 months ago
Motion planning under uncertainty for robotic tasks with long time horizons
Abstract Partially observable Markov decision processes (POMDPs) are a principled mathematical framework for planning under uncertainty, a crucial capability for reliable operation...
Hanna Kurniawati, Yanzhu Du, David Hsu, Wee Sun Le...
LION
2007
Springer
192views Optimization» more  LION 2007»
14 years 1 months ago
Learning While Optimizing an Unknown Fitness Surface
This paper is about Reinforcement Learning (RL) applied to online parameter tuning in Stochastic Local Search (SLS) methods. In particular a novel application of RL is considered i...
Roberto Battiti, Mauro Brunato, Paolo Campigotto
NIPS
1998
13 years 8 months ago
Risk Sensitive Reinforcement Learning
In this paper, we consider Markov Decision Processes (MDPs) with error states. Error states are those states entering which is undesirable or dangerous. We define the risk with re...
Ralph Neuneier, Oliver Mihatsch