Boosting algorithm is understood as the gradient descent algorithm of a loss function. It is often pointed out that the typical boosting algorithm, Adaboost, is seriously affected ...
Ridge regression is a well established method to shrink regression parameters towards zero, thereby securing existence of estimates. The present paper investigates several approac...
We develop and evaluate an approach to causal modeling based on time series data, collectively referred to as“grouped graphical Granger modeling methods.” Graphical Granger mo...
Aurelie C. Lozano, Naoki Abe, Yan Liu, Saharon Ros...
A novel technique is presented to construct sparse regression models based on the orthogonal least square method with boosting. This technique tunes the mean vector and diagonal c...
In a seminal paper, Amari (1998) proved that learning can be made more efficient when one uses the intrinsic Riemannian structure of the algorithms' spaces of parameters to po...