Sciweavers

Share
ICML
2004
IEEE

Learning Bayesian network classifiers by maximizing conditional likelihood

12 years 2 months ago
Learning Bayesian network classifiers by maximizing conditional likelihood
Bayesian networks are a powerful probabilistic representation, and their use for classification has received considerable attention. However, they tend to perform poorly when learned in the standard way. This is attributable to a mismatch between the objective function used (likelihood or a function thereof) and the goal of classification (maximizing accuracy or conditional likelihood). Unfortunately, the computational cost of optimizing structure and parameters for conditional likelihood is prohibitive. In this paper we show that a simple approximation-choosing structures by maximizing conditional likelihood while setting parameters by maximum likelihood--yields good results. On a large suite of benchmark datasets, this approach produces better class probability estimates than naive Bayes, TAN, and generatively-trained Bayesian networks.
Daniel Grossman, Pedro Domingos
Added 17 Nov 2009
Updated 17 Nov 2009
Type Conference
Year 2004
Where ICML
Authors Daniel Grossman, Pedro Domingos
Comments (0)
books