Sciweavers

100 search results - page 4 / 20
» Distributed training of large scale exponential language mod...
Sort
View
HIPC
2005
Springer
15 years 3 months ago
The Impact of Noise on the Scaling of Collectives: A Theoretical Approach
The performance of parallel applications running on large clusters is known to degrade due to the interference of kernel and daemon activities on individual nodes, often referred t...
Saurabh Agarwal, Rahul Garg, Nisheeth K. Vishnoi
NIPS
2000
14 years 11 months ago
A Neural Probabilistic Language Model
A goal of statistical language modeling is to learn the joint probability function of sequences of words in a language. This is intrinsically difficult because of the curse of dim...
Yoshua Bengio, Réjean Ducharme, Pascal Vinc...
EMNLP
2010
14 years 7 months ago
Training Continuous Space Language Models: Some Practical Issues
Using multi-layer neural networks to estimate the probabilities of word sequences is a promising research area in statistical language modeling, with applications in speech recogn...
Hai Son Le, Alexandre Allauzen, Guillaume Wisniews...
ICML
2004
IEEE
15 years 10 months ago
Training conditional random fields via gradient tree boosting
Conditional Random Fields (CRFs; Lafferty, McCallum, & Pereira, 2001) provide a flexible and powerful model for learning to assign labels to elements of sequences in such appl...
Thomas G. Dietterich, Adam Ashenfelter, Yaroslav B...
87
Voted
PKDD
2010
Springer
178views Data Mining» more  PKDD 2010»
14 years 8 months ago
Large-Scale Support Vector Learning with Structural Kernels
Abstract. In this paper, we present an extensive study of the cuttingplane algorithm (CPA) applied to structural kernels for advanced text classification on large datasets. In par...
Aliaksei Severyn, Alessandro Moschitti