Sciweavers

ACL
2001

Multi-Class Composite N-gram Language Model for Spoken Language Processing Using Multiple Word Clusters

13 years 5 months ago
Multi-Class Composite N-gram Language Model for Spoken Language Processing Using Multiple Word Clusters
In this paper, a new language model, the Multi-Class Composite N-gram, is proposed to avoid a data sparseness problem for spoken language in that it is difficult to collect training data. The Multi-Class Composite N-gram maintains an accurate word prediction capability and reliability for sparse data with a compact model size based on multiple word clusters, called MultiClasses. In the Multi-Class, the statistical connectivity at each position of the N-grams is regarded as word attributes, and one word cluster each is created to represent the positional attributes. Furthermore, by introducing higher order word N-grams through the grouping of frequent word successions, Multi-Class N-grams are extended to Multi-Class Composite N-grams. In experiments, the Multi-Class Composite N-grams result in 9.5% lower perplexity and a 16% lower word error rate in speech recognition with a 40% smaller parameter size than conventional word 3-grams.
Hirofumi Yamamoto, Shuntaro Isogai, Yoshinori Sagi
Added 31 Oct 2010
Updated 31 Oct 2010
Type Conference
Year 2001
Where ACL
Authors Hirofumi Yamamoto, Shuntaro Isogai, Yoshinori Sagisaka
Comments (0)