Sciweavers

NAACL
2010

Painless Unsupervised Learning with Features

13 years 2 months ago
Painless Unsupervised Learning with Features
We show how features can easily be added to standard generative models for unsupervised learning, without requiring complex new training methods. In particular, each component multinomial of a generative model can be turned into a miniature logistic regression model if feature locality permits. The intuitive EM algorithm still applies, but with a gradient-based M-step familiar from discriminative training of logistic regression models. We apply this technique to part-of-speech induction, grammar induction, word alignment, and word segmentation, incorporating a few linguistically-motivated features into the standard generative model for each task. These feature-enhanced models each outperform their basic counterparts by a substantial margin, and even compete with and surpass more complex state-of-the-art models.
Taylor Berg-Kirkpatrick, Alexandre Bouchard-C&ocir
Added 14 Feb 2011
Updated 14 Feb 2011
Type Journal
Year 2010
Where NAACL
Authors Taylor Berg-Kirkpatrick, Alexandre Bouchard-Côté, John DeNero, Dan Klein
Comments (0)