Sciweavers

40 search results - page 4 / 8
» Markov decision process (MDP) framework for optimizing softw...
Sort
View
AAAI
2012
11 years 9 months ago
A Dynamic Rationalization of Distance Rationalizability
Distance rationalizability is an intuitive paradigm for developing and studying voting rules: given a notion of consensus and a distance function on preference profiles, a ration...
Craig Boutilier, Ariel D. Procaccia
RSS
2007
136views Robotics» more  RSS 2007»
13 years 8 months ago
The Stochastic Motion Roadmap: A Sampling Framework for Planning with Markov Motion Uncertainty
— We present a new motion planning framework that explicitly considers uncertainty in robot motion to maximize the probability of avoiding collisions and successfully reaching a ...
Ron Alterovitz, Thierry Siméon, Kenneth Y. ...
JSAC
2010
107views more  JSAC 2010»
13 years 5 months ago
Online learning in autonomic multi-hop wireless networks for transmitting mission-critical applications
Abstract—In this paper, we study how to optimize the transmission decisions of nodes aimed at supporting mission-critical applications, such as surveillance, security monitoring,...
Hsien-Po Shiang, Mihaela van der Schaar
ICML
2007
IEEE
14 years 8 months ago
Multi-task reinforcement learning: a hierarchical Bayesian approach
We consider the problem of multi-task reinforcement learning, where the agent needs to solve a sequence of Markov Decision Processes (MDPs) chosen randomly from a fixed but unknow...
Aaron Wilson, Alan Fern, Soumya Ray, Prasad Tadepa...
NIPS
2008
13 years 8 months ago
MDPs with Non-Deterministic Policies
Markov Decision Processes (MDPs) have been extensively studied and used in the context of planning and decision-making, and many methods exist to find the optimal policy for probl...
Mahdi Milani Fard, Joelle Pineau