Sciweavers

155 search results - page 21 / 31
» Domain-Theoretic Formulation of Linear Boundary Value Proble...
Sort
View
AAAI
1997
13 years 9 months ago
Incremental Methods for Computing Bounds in Partially Observable Markov Decision Processes
Partially observable Markov decision processes (POMDPs) allow one to model complex dynamic decision or control problems that include both action outcome uncertainty and imperfect ...
Milos Hauskrecht
JMLR
2006
150views more  JMLR 2006»
13 years 7 months ago
Exact 1-Norm Support Vector Machines Via Unconstrained Convex Differentiable Minimization
Support vector machines utilizing the 1-norm, typically set up as linear programs (Mangasarian, 2000; Bradley and Mangasarian, 1998), are formulated here as a completely unconstra...
Olvi L. Mangasarian
ML
2002
ACM
168views Machine Learning» more  ML 2002»
13 years 7 months ago
On Average Versus Discounted Reward Temporal-Difference Learning
We provide an analytical comparison between discounted and average reward temporal-difference (TD) learning with linearly parameterized approximations. We first consider the asympt...
John N. Tsitsiklis, Benjamin Van Roy
AUTOMATICA
1999
119views more  AUTOMATICA 1999»
13 years 7 months ago
Robust L2-gain control for nonlinear systems with projection dynamics and input constraints: an example from traffic control
We formulate the L2-gain control problem for a general nonlinear, state-space system with projection dynamics in the state evolution and hard constraints on the set of admissible ...
Joseph A. Ball, Martin V. Day, Tungsheng Yu, Pushk...
SIAMCO
2002
121views more  SIAMCO 2002»
13 years 7 months ago
Consistent Approximations and Approximate Functions and Gradients in Optimal Control
As shown in [7], optimal control problems with either ODE or PDE dynamics can be solved efficiently using a setting of consistent approximations obtained by numerical discretizati...
Olivier Pironneau, Elijah Polak