Sciweavers

109 search results - page 3 / 22
» Algorithm Selection as a Bandit Problem with Unbounded Losse...
Sort
View
ICDM
2007
IEEE
138views Data Mining» more  ICDM 2007»
13 years 10 months ago
Bandit-Based Algorithms for Budgeted Learning
We explore the problem of budgeted machine learning, in which the learning algorithm has free access to the training examples’ labels but has to pay for each attribute that is s...
Kun Deng, Chris Bourke, Stephen D. Scott, Julie Su...
ICML
2010
IEEE
13 years 5 months ago
Efficient Selection of Multiple Bandit Arms: Theory and Practice
We consider the general, widely applicable problem of selecting from n real-valued random variables a subset of size m of those with the highest means, based on as few samples as ...
Shivaram Kalyanakrishnan, Peter Stone
GECCO
2010
Springer
193views Optimization» more  GECCO 2010»
13 years 9 months ago
Fitness-AUC bandit adaptive strategy selection vs. the probability matching one within differential evolution: an empirical comp
The choice of which of the available strategies should be used within the Differential Evolution algorithm for a given problem is not trivial, besides being problem-dependent and...
Álvaro Fialho, Marc Schoenauer, Michè...
NIPS
2008
13 years 6 months ago
Algorithms for Infinitely Many-Armed Bandits
We consider multi-armed bandit problems where the number of arms is larger than the possible number of experiments. We make a stochastic assumption on the mean-reward of a new sel...
Yizao Wang, Jean-Yves Audibert, Rémi Munos
COLT
2005
Springer
13 years 6 months ago
From External to Internal Regret
External regret compares the performance of an online algorithm, selecting among N actions, to the performance of the best of those actions in hindsight. Internal regret compares ...
Avrim Blum, Yishay Mansour