In this paper, we propose a new method, Parametric Embedding (PE), for visualizing the posteriors estimated over a mixture model. PE simultaneously embeds both objects and their c...
Tomoharu Iwata, Kazumi Saito, Naonori Ueda, Sean S...
We combine the strengths of Bayesian modeling and synchronous grammar in unsupervised learning of basic translation phrase pairs. The structured space of a synchronous grammar is ...
Hao Zhang, Chris Quirk, Robert C. Moore, Daniel Gi...
In this paper, we propose a novel method for semi-supervised learning of nonprojective log-linear dependency parsers using directly expressed linguistic prior knowledge (e.g. a no...
This paper presents a new dependence language modeling approach to information retrieval. The approach extends the basic language modeling approach based on unigram by relaxing th...
Jianfeng Gao, Jian-Yun Nie, Guangyuan Wu, Guihong ...
Statistical language models estimate the probability of a word occurring in a given context. The most common language models rely on a discrete enumeration of predictive contexts ...
John Blitzer, Kilian Q. Weinberger, Lawrence K. Sa...