Sciweavers

JSTSP
2016

Stochastic Spectral Descent for Discrete Graphical Models

8 years 14 days ago
Stochastic Spectral Descent for Discrete Graphical Models
—Interest in deep probabilistic graphical models has increased in recent years, due to their state-of-the-art performance on many machine learning applications. Such models are typically trained with the stochastic gradient method, which can take a significant number of iterations to converge. Since the computational cost of gradient estimation is prohibitive even for modestly-sized models, training becomes slow and practicallyusable models are kept small. In this paper we propose a new, largely tuning-free algorithm to address this problem. Our approach derives novel majorization bounds based on the Schatten-∞ norm. Intriguingly, the minimizers of these bounds can be interpreted as gradient methods in a non-Euclidean space. We thus propose using a stochastic gradient method in non-Euclidean space. We both provide simple conditions under which our algorithm is guaranteed to converge, and demonstrate empirically that our algorithm leads to dramatically faster training and improved ...
David E. Carlson, Ya-Ping Hsieh, Edo Collins, Lawr
Added 07 Apr 2016
Updated 07 Apr 2016
Type Journal
Year 2016
Where JSTSP
Authors David E. Carlson, Ya-Ping Hsieh, Edo Collins, Lawrence Carin, Volkan Cevher
Comments (0)