Sciweavers

109 search results - page 13 / 22
» Model Checking Markov Reward Models with Impulse Rewards
Sort
View
AAAI
1997
13 years 11 months ago
Incremental Methods for Computing Bounds in Partially Observable Markov Decision Processes
Partially observable Markov decision processes (POMDPs) allow one to model complex dynamic decision or control problems that include both action outcome uncertainty and imperfect ...
Milos Hauskrecht
CDC
2009
IEEE
169views Control Systems» more  CDC 2009»
14 years 3 months ago
Parametric regret in uncertain Markov decision processes
— We consider decision making in a Markovian setup where the reward parameters are not known in advance. Our performance criterion is the gap between the performance of the best ...
Huan Xu, Shie Mannor
COMPSAC
2009
IEEE
14 years 3 months ago
Modeling and Predicting Software Failure Costs
—For software, the costs of failures are not clearly understood. Often, these costs disappear in the costs of testing, the general developments costs, or the operating expenses. ...
Michael Grottke, Christian A. Graf
ILP
2007
Springer
14 years 4 months ago
Building Relational World Models for Reinforcement Learning
Abstract. Many reinforcement learning domains are highly relational. While traditional temporal-difference methods can be applied to these domains, they are limited in their capaci...
Trevor Walker, Lisa Torrey, Jude W. Shavlik, Richa...
CORR
2007
Springer
143views Education» more  CORR 2007»
13 years 10 months ago
On Myopic Sensing for Multi-Channel Opportunistic Access
We consider a multi-channel opportunistic communication system where the states of these channels evolve as independent and statistically identical Markov chains (the Gilbert-Elli...
Qing Zhao, Bhaskar Krishnamachari, Keqin Liu