Sciweavers

350 search results - page 40 / 70
» Complexity of Planning with Partial Observability
Sort
View
ICANN
2007
Springer
14 years 2 months ago
Solving Deep Memory POMDPs with Recurrent Policy Gradients
Abstract. This paper presents Recurrent Policy Gradients, a modelfree reinforcement learning (RL) method creating limited-memory stochastic policies for partially observable Markov...
Daan Wierstra, Alexander Förster, Jan Peters,...
AINA
2007
IEEE
13 years 12 months ago
Domain Modelling for Ubiquitous Computing Applications
Many Ubiquitous computing applications can be considered as planning and acting problems in environments characterised by uncertainty and partial observability. Such systems rely ...
Anthony Harrington, Vinny Cahill
AAAI
2012
11 years 10 months ago
Tree-Based Solution Methods for Multiagent POMDPs with Delayed Communication
Planning under uncertainty is an important and challenging problem in multiagent systems. Multiagent Partially Observable Markov Decision Processes (MPOMDPs) provide a powerful fr...
Frans Adriaan Oliehoek, Matthijs T. J. Spaan
ICASSP
2008
IEEE
14 years 2 months ago
Bayesian update of dialogue state for robust dialogue systems
This paper presents a new framework for accumulating beliefs in spoken dialogue systems. The technique is based on updating a Bayesian Network that represents the underlying state...
Blaise Thomson, Jost Schatzmann, Steve Young
TACAS
2007
Springer
116views Algorithms» more  TACAS 2007»
14 years 1 months ago
Model Checking on Trees with Path Equivalences
For specifying and verifying branching-time requirements, a reactive system is traditionally modeled as a labeled tree, where a path in the tree encodes a possible execution of the...
Rajeev Alur, Pavol Cerný, Swarat Chaudhuri