In apprenticeship learning, the goal is to learn a policy in a Markov decision process that is at least as good as a policy demonstrated by an expert. The difficulty arises in tha...
Partially Observable Markov Decision Processes (POMDPs) have succeeded in planning domains that require balancing actions that increase an agent's knowledge and actions that ...
Abstract. The Factored Markov Decision Process (FMDP) framework is a standard representation for sequential decision problems under uncertainty where the state is represented as a ...
Olga Kozlova, Olivier Sigaud, Pierre-Henri Wuillem...
The Shadow Box is a tangible computing project that exploits visual association and auditory clues to teach children the representational relationship between words and their mean...
Ja-Young Sung, Aaron Levisohn, Ji-won Song, Ben To...
Many applications in information retrieval, natural language processing, data mining, and related fields require a ranking of instances with respect to a specified criteria as op...