Sciweavers

458 search results - page 80 / 92
» Q-Decomposition for Reinforcement Learning Agents
Sort
View
AAAI
2008
13 years 10 months ago
Adaptive Importance Sampling with Automatic Model Selection in Value Function Approximation
Off-policy reinforcement learning is aimed at efficiently reusing data samples gathered in the past, which is an essential problem for physically grounded AI as experiments are us...
Hirotaka Hachiya, Takayuki Akiyama, Masashi Sugiya...
SASO
2008
IEEE
14 years 2 months ago
Self-Adaptive Dissemination of Data in Dynamic Sensor Networks
The distribution of data in large dynamic wireless sensor networks presents a difficult problem due to node mobility, link failures, and traffic congestion. In this paper, we pr...
David Dorsey, Bjorn Jay Carandang, Moshe Kam, Chri...
NIPS
2004
13 years 9 months ago
Responding to Modalities with Different Latencies
Motor control depends on sensory feedback in multiple modalities with different latencies. In this paper we consider within the framework of reinforcement learning how different s...
Fredrik Bissmarck, Hiroyuki Nakahara, Kenji Doya, ...
IIE
2007
63views more  IIE 2007»
13 years 7 months ago
Investigation of Q-Learning in the Context of a Virtual Learning Environment
We investigate the possibility to apply a known machine learning algorithm of Q-learning in the domain of a Virtual Learning Environment (VLE). It is important in this problem doma...
Dalia Baziukaite
CCIA
2009
Springer
13 years 8 months ago
Interaction, observance or both? Study of the effects on convention emergence
Abstract. Social conventions are useful self-sustaining protocols for groups to coordinate behavior without a centralized entity enforcing coordination. The emergence of such conve...
Daniel Villatoro, Jordi Sabater-Mir, Sandip Sen