Sciweavers

4446 search results - page 7 / 890
» Learning Observer Agents
Sort
View
ECML
2005
Springer
14 years 28 days ago
Using Rewards for Belief State Updates in Partially Observable Markov Decision Processes
Partially Observable Markov Decision Processes (POMDP) provide a standard framework for sequential decision making in stochastic environments. In this setting, an agent takes actio...
Masoumeh T. Izadi, Doina Precup
ICMLA
2010
13 years 5 months ago
Multi-Agent Inverse Reinforcement Learning
Learning the reward function of an agent by observing its behavior is termed inverse reinforcement learning and has applications in learning from demonstration or apprenticeship l...
Sriraam Natarajan, Gautam Kunapuli, Kshitij Judah,...
IDEAL
2004
Springer
14 years 23 days ago
Generating and Applying Rules for Interval Valued Fuzzy Observations
Abstract. One of the objectives of intelligent data engineering and automated learning is to develop algorithms that learn the environment, generate rules, and take possible course...
André de Korvin, Chenyi Hu, Ping Chen
ECAI
2004
Springer
14 years 24 days ago
Combining Multiple Answers for Learning Mathematical Structures from Visual Observation
Learning general truths from the observation of simple domains and, further, learning how to use this knowledge are essential capabilities for any intelligent agent to understand ...
Paulo Santos, Derek R. Magee, Anthony G. Cohn, Dav...
ICML
1995
IEEE
14 years 8 months ago
Learning Policies for Partially Observable Environments: Scaling Up
Partially observable Markov decision processes (pomdp's) model decision problems in which an agent tries to maximize its reward in the face of limited and/or noisy sensor fee...
Michael L. Littman, Anthony R. Cassandra, Leslie P...