We address the problem of learning the mapping between words and their possible pronunciations in terms of sub-word units. Most previous approaches have involved generative modeli...
We consider boosting algorithms that maintain a distribution over a set of examples. At each iteration a weak hypothesis is received and the distribution is updated. We motivate t...
We propose a new family of latent variable models called max-margin min-entropy (m3e) models, which define a distribution over the output and the hidden variables conditioned on ...
Kevin Miller, M. Pawan Kumar, Benjamin Packer, Dan...
Boosting methods are known not to usually overfit training data even as the size of the generated classifiers becomes large. Schapire et al. attempted to explain this phenomenon i...
Generalization bounds depending on the margin of a classifier are a relatively recent development. They provide an explanation of the performance of state-of-the-art learning syste...