Sciweavers

499 search results - page 26 / 100
» Model Minimization in Markov Decision Processes
Sort
View
ICASSP
2009
IEEE
15 years 11 months ago
Combining mixture weight pruning and quantization for small-footprint speech recognition
Semi-continuous acoustic models, where the output distributions for all Hidden Markov Model states share a common codebook of Gaussian density functions, are a well-known and prov...
David Huggins-Daines, Alexander I. Rudnicky

Publication
233views
14 years 3 months ago
Sparse reward processes
We introduce a class of learning problems where the agent is presented with a series of tasks. Intuitively, if there is relation among those tasks, then the information gained duri...
Christos Dimitrakakis
ATAL
2007
Springer
15 years 8 months ago
Interactive dynamic influence diagrams
This paper extends the framework of dynamic influence diagrams (DIDs) to the multi-agent setting. DIDs are computational representations of the Partially Observable Markov Decisio...
Kyle Polich, Piotr J. Gmytrasiewicz
CORR
2010
Springer
105views Education» more  CORR 2010»
15 years 3 months ago
Optimism in Reinforcement Learning Based on Kullback-Leibler Divergence
We consider model-based reinforcement learning in finite Markov Decision Processes (MDPs), focussing on so-called optimistic strategies. Optimism is usually implemented by carryin...
Sarah Filippi, Olivier Cappé, Aurelien Gari...
AAAI
2010
15 years 5 months ago
Representation Discovery in Sequential Decision Making
Automatically constructing novel representations of tasks from analysis of state spaces is a longstanding fundamental challenge in AI. I review recent progress on this problem for...
Sridhar Mahadevan