We apply Stochastic Meta-Descent (SMD), a stochastic gradient optimization method with gain vector adaptation, to the training of Conditional Random Fields (CRFs). On several larg...
S. V. N. Vishwanathan, Nicol N. Schraudolph, Mark ...
Over the past few years, a number of approximate inference algorithms for networked data have been put forth. We empirically compare the performance of three of the popular algori...
We develop latent Dirichlet allocation with WORDNET (LDAWN), an unsupervised probabilistic topic model that includes word sense as a hidden variable. We develop a probabilistic po...
We present and derive a new stick-breaking construction of the beta process. The construction is closely related to a special case of the stick-breaking construction of the Dirich...
John William Paisley, Aimee Zaas, Christopher W. W...
This paper presents a technique for learning parameterized implied constraints. They can be added to a model to improve the solving process. Experiments on implied Gcc constraints ...