Sciweavers

ICML
2006
IEEE

Full Bayesian network classifiers

14 years 5 months ago
Full Bayesian network classifiers
The structure of a Bayesian network (BN) encodes variable independence. Learning the structure of a BN, however, is typically of high computational complexity. In this paper, we explore and represent variable independence in learning conditional probability tables (CPTs), instead of in learning structure. A full Bayesian network is used as the structure and a decision tree is learned for each CPT. The resulting model is called full Bayesian network classifiers (FBCs). In learning an FBC, learning the decision trees for CPTs captures essentially both variable independence and context-specific independence. We present a novel, efficient decision tree learning, which is also effective in the context of FBC learning. In our experiments, the FBC learning algorithm demonstrates better performance in both classification and ranking compared with other stateof-the-art learning algorithms. In addition, its reduced effort on structure learning makes its time complexity quite low as well.
Jiang Su, Harry Zhang
Added 17 Nov 2009
Updated 17 Nov 2009
Type Conference
Year 2006
Where ICML
Authors Jiang Su, Harry Zhang
Comments (0)