— Many deterministic algorithms in the context of constrained optimization require the first-order derivatives, or the gradient vectors, of the objective and constraint function...
Stephanus Daniel Handoko, Chee Keong Kwoh, Yew-Soo...
Stochastic gradient descent (SGD) uses approximate gradients estimated from subsets of the training data and updates the parameters in an online fashion. This learning framework i...
We describe a distributed system for texture mapping implicit surfaces. The method uses a particle system associated with the gradient vector field of the function that defines an...
Ruben Zonenschein, Jonas Gomes, Luiz Velho, Noemi ...
On large datasets, the popular training approach has been stochastic gradient descent (SGD). This paper proposes a modification of SGD, called averaged SGD with feedback (ASF), tha...
A new class of parallel normalized preconditioned conjugate gradient type methods in conjunction with normalized approximate inverses algorithms, based on normalized approximate f...