Sciweavers

231 search results - page 16 / 47
» Active Learning in Partially Observable Markov Decision Proc...
Sort
View
CORR
2008
Springer
103views Education» more  CORR 2008»
13 years 8 months ago
Quickest Change Detection of a Markov Process Across a Sensor Array
Recent attention in quickest change detection in the multi-sensor setting has been on the case where the densities of the observations change at the same instant at all the sensor...
Vasanthan Raghavan, Venugopal V. Veeravalli
GLOBECOM
2010
IEEE
13 years 6 months ago
Maximize Secondary User Throughput via Optimal Sensing in Multi-Channel Cognitive Radio Networks
In a cognitive radio network, the full-spectrum is usually divided into multiple channels. However, due to the hardware and energy constraints, a cognitive user (also called second...
Shimin Gong, Ping Wang, Wei Liu, Wei Yuan
ATAL
2009
Springer
14 years 3 months ago
SarsaLandmark: an algorithm for learning in POMDPs with landmarks
Reinforcement learning algorithms that use eligibility traces, such as Sarsa(λ), have been empirically shown to be effective in learning good estimated-state-based policies in pa...
Michael R. James, Satinder P. Singh
NIPS
2008
13 years 10 months ago
Particle Filter-based Policy Gradient in POMDPs
Our setting is a Partially Observable Markov Decision Process with continuous state, observation and action spaces. Decisions are based on a Particle Filter for estimating the bel...
Pierre-Arnaud Coquelin, Romain Deguest, Rém...
AIPS
2003
13 years 10 months ago
Synthesis of Hierarchical Finite-State Controllers for POMDPs
We develop a hierarchical approach to planning for partially observable Markov decision processes (POMDPs) in which a policy is represented as a hierarchical finite-state control...
Eric A. Hansen, Rong Zhou