Sciweavers

499 search results - page 26 / 100
» Model Minimization in Markov Decision Processes
Sort
View
ICASSP
2009
IEEE
14 years 3 months ago
Combining mixture weight pruning and quantization for small-footprint speech recognition
Semi-continuous acoustic models, where the output distributions for all Hidden Markov Model states share a common codebook of Gaussian density functions, are a well-known and prov...
David Huggins-Daines, Alexander I. Rudnicky

Publication
233views
12 years 7 months ago
Sparse reward processes
We introduce a class of learning problems where the agent is presented with a series of tasks. Intuitively, if there is relation among those tasks, then the information gained duri...
Christos Dimitrakakis
ATAL
2007
Springer
14 years 21 days ago
Interactive dynamic influence diagrams
This paper extends the framework of dynamic influence diagrams (DIDs) to the multi-agent setting. DIDs are computational representations of the Partially Observable Markov Decisio...
Kyle Polich, Piotr J. Gmytrasiewicz
CORR
2010
Springer
105views Education» more  CORR 2010»
13 years 7 months ago
Optimism in Reinforcement Learning Based on Kullback-Leibler Divergence
We consider model-based reinforcement learning in finite Markov Decision Processes (MDPs), focussing on so-called optimistic strategies. Optimism is usually implemented by carryin...
Sarah Filippi, Olivier Cappé, Aurelien Gari...
AAAI
2010
13 years 10 months ago
Representation Discovery in Sequential Decision Making
Automatically constructing novel representations of tasks from analysis of state spaces is a longstanding fundamental challenge in AI. I review recent progress on this problem for...
Sridhar Mahadevan