Sciweavers

Share
ICDM
2007
IEEE

Bandit-Based Algorithms for Budgeted Learning

11 years 12 months ago
Bandit-Based Algorithms for Budgeted Learning
We explore the problem of budgeted machine learning, in which the learning algorithm has free access to the training examples’ labels but has to pay for each attribute that is specified. This learning model is appropriate in many areas, including medical applications. We present new algorithms for choosing which attributes to purchase of which examples in the budgeted learning model based on algorithms for the multi-armed bandit problem. All of our approaches outperformed the current state of the art. Furthermore, we present a new means for selecting an example to purchase after the attribute is selected, instead of selecting an example uniformly at random, which is typically done. Our new example selection method improved performance of all the algorithms we tested, both ours and those in the literature.
Kun Deng, Chris Bourke, Stephen D. Scott, Julie Su
Added 03 Jun 2010
Updated 03 Jun 2010
Type Conference
Year 2007
Where ICDM
Authors Kun Deng, Chris Bourke, Stephen D. Scott, Julie Sunderman, Yaling Zheng
Comments (0)
books