A situation where training and test samples follow different input distributions is called covariate shift. Under covariate shift, standard learning methods such as maximum likeli...
Recent research has studied the role of sparsity in high dimensional regression and signal reconstruction, establishing theoretical limits for recovering sparse models from sparse...
Shuheng Zhou, John D. Lafferty, Larry A. Wasserman
Current computational models of bottom-up and top-down components of attention are predictive of eye movements across a range of stimuli and of simple, fixed visual tasks (such a...
We present an efficient generalization of the sparse pseudo-input Gaussian process (SPGP) model developed by Snelson and Ghahramani [1], applying it to binary classification pro...
Brain-Computer Interfaces can suffer from a large variance of the subject conditions within and across sessions. For example vigilance fluctuations in the individual, variable ta...
Benjamin Blankertz, Motoaki Kawanabe, Ryota Tomiok...
We present a probabilistic approach to language change in which word forms are represented by phoneme sequences that undergo stochastic edits along the branches of a phylogenetic ...
Content-based image suggestion (CBIS) targets the recommendation of products based on user preferences on the visual content of images. In this paper, we motivate both feature sel...
Neural spike trains present challenges to analytical efforts due to their noisy, spiking nature. Many studies of neuroscientific and neural prosthetic importance rely on a smooth...
John P. Cunningham, Byron M. Yu, Krishna V. Shenoy...