Sciweavers

98 search results - page 7 / 20
» Using Rewards for Belief State Updates in Partially Observab...
Sort
View
AAAI
2010
13 years 9 months ago
Compressing POMDPs Using Locality Preserving Non-Negative Matrix Factorization
Partially Observable Markov Decision Processes (POMDPs) are a well-established and rigorous framework for sequential decision-making under uncertainty. POMDPs are well-known to be...
Georgios Theocharous, Sridhar Mahadevan
AAAI
2000
13 years 9 months ago
Back to the Future for Consistency-Based Trajectory Tracking
Given a model of a physical process and a sequence of commands and observations received over time, the task of an autonomous controller is to determine the likely states of the p...
James Kurien, P. Pandurang Nayak
ICML
2004
IEEE
14 years 8 months ago
Utile distinction hidden Markov models
This paper addresses the problem of constructing good action selection policies for agents acting in partially observable environments, a class of problems generally known as Part...
Daan Wierstra, Marco Wiering
CDC
2008
IEEE
140views Control Systems» more  CDC 2008»
14 years 2 months ago
Information state for Markov decision processes with network delays
We consider a networked control system, where each subsystem evolves as a Markov decision process (MDP). Each subsystem is coupled to its neighbors via communication links over wh...
Sachin Adlakha, Sanjay Lall, Andrea J. Goldsmith
HICSS
2003
IEEE
123views Biometrics» more  HICSS 2003»
14 years 27 days ago
Issues in Rational Planning in Multi-Agent Settings
We adopt the decision-theoretic principle of expected utility maximization as a paradigm for designing autonomous rational agents operating in multi-agent environments. We use the...
Piotr J. Gmytrasiewicz