Sciweavers

JMLR
2002

Machine Learning with Data Dependent Hypothesis Classes

13 years 4 months ago
Machine Learning with Data Dependent Hypothesis Classes
We extend the VC theory of statistical learning to data dependent spaces of classifiers. This theory can be viewed as a decomposition of classifier design into two components; the first component is a restriction to a data dependent hypothesis class and the second is empirical risk minimization within that class. We define a measure of complexity for data dependent hypothesis classes and provide data dependent versions of bounds on error deviance and estimation error. We also provide a structural risk minimization procedure over data dependent hierarchies and prove consistency. We use this theory to provide a framework for studying the trade-offs between performance and computational complexity in classifier design. As a consequence we obtain a new family of classifiers with dimension independent performance bounds and efficient learning procedures.
Adam Cannon, J. Mark Ettinger, Don R. Hush, Clint
Added 22 Dec 2010
Updated 22 Dec 2010
Type Journal
Year 2002
Where JMLR
Authors Adam Cannon, J. Mark Ettinger, Don R. Hush, Clint Scovel
Comments (0)