Sciweavers

Share
ICMLA
2008

Decision Tree Ensemble: Small Heterogeneous Is Better Than Large Homogeneous

9 years 4 months ago
Decision Tree Ensemble: Small Heterogeneous Is Better Than Large Homogeneous
Using decision trees that split on randomly selected attributes is one way to increase the diversity within an ensemble of decision trees. Another approach increases diversity by combining multiple tree algorithms. The random forest approach has become popular because it is simple and yields good results with common datasets. We present a technique that combines heterogeneous tree algorithms and contrast it with homogeneous forest algorithms. Our results indicate that random forests do poorly when faced with irrelevant attributes, while our heterogeneous technique handles them robustly. Further, we show that large ensembles of random trees are more susceptible to diminishing returns than our technique. We are able to obtain better results across a large number of common datasets with a significantly smaller ensemble.
Michael Gashler, Christophe G. Giraud-Carrier, Ton
Added 29 Oct 2010
Updated 29 Oct 2010
Type Conference
Year 2008
Where ICMLA
Authors Michael Gashler, Christophe G. Giraud-Carrier, Tony R. Martinez
Comments (0)
books