Sciweavers

6 search results - page 1 / 2
» Tuning Bandit Algorithms in Stochastic Environments
Sort
View
ALT
2007
Springer
14 years 1 months ago
Tuning Bandit Algorithms in Stochastic Environments
Algorithms based on upper-confidence bounds for balancing exploration and exploitation are gaining popularity since they are easy to implement, efficient and effective. In this p...
Jean-Yves Audibert, Rémi Munos, Csaba Szepe...
EVOW
2012
Springer
12 years 15 days ago
Hyperparameter Tuning in Bandit-Based Adaptive Operator Selection
We are using bandit-based adaptive operator selection while autotuning parallel computer programs. The autotuning, which uses evolutionary algorithm-based stochastic sampling, take...
Maciej Pacula, Jason Ansel, Saman P. Amarasinghe, ...
COLT
2008
Springer
13 years 6 months ago
Adapting to a Changing Environment: the Brownian Restless Bandits
In the multi-armed bandit (MAB) problem there are k distributions associated with the rewards of playing each of k strategies (slot machine arms). The reward distributions are ini...
Aleksandrs Slivkins, Eli Upfal
COLT
2010
Springer
13 years 2 months ago
Best Arm Identification in Multi-Armed Bandits
We consider the problem of finding the best arm in a stochastic multi-armed bandit game. The regret of a forecaster is here defined by the gap between the mean reward of the optim...
Jean-Yves Audibert, Sébastien Bubeck, R&eac...
ATAL
2008
Springer
13 years 6 months ago
An approach to online optimization of heuristic coordination algorithms
Due to computational intractability, large scale coordination algorithms are necessarily heuristic and hence require tuning for particular environments. In domains where character...
Jumpol Polvichai, Paul Scerri, Michael Lewis