Sciweavers

17 search results - page 2 / 4
» The Budgeted Multi-armed Bandit Problem
Sort
View
CORR
2006
Springer
83views Education» more  CORR 2006»
14 years 9 months ago
How to Beat the Adaptive Multi-Armed Bandit
The multi-armed bandit is a concise model for the problem of iterated decision-making under uncertainty. In each round, a gambler must pull one of K arms of a slot machine, withou...
Varsha Dani, Thomas P. Hayes
CORR
2010
Springer
143views Education» more  CORR 2010»
14 years 6 months ago
The Non-Bayesian Restless Multi-Armed Bandit: a Case of Near-Logarithmic Regret
In the classic Bayesian restless multi-armed bandit (RMAB) problem, there are N arms, with rewards on all arms evolving at each time as Markov chains with known parameters. A play...
Wenhan Dai, Yi Gai, Bhaskar Krishnamachari, Qing Z...
96
Voted
CORR
2010
Springer
187views Education» more  CORR 2010»
14 years 9 months ago
Learning in A Changing World: Non-Bayesian Restless Multi-Armed Bandit
We consider the restless multi-armed bandit (RMAB) problem with unknown dynamics. In this problem, at each time, a player chooses K out of N (N > K) arms to play. The state of ...
Haoyang Liu, Keqin Liu, Qing Zhao
CORR
2010
Springer
152views Education» more  CORR 2010»
14 years 4 months ago
Combinatorial Network Optimization with Unknown Variables: Multi-Armed Bandits with Linear Rewards
In the classic multi-armed bandits problem, the goal is to have a policy for dynamically operating arms that each yield stochastic rewards with unknown means. The key metric of int...
Yi Gai, Bhaskar Krishnamachari, Rahul Jain
96
Voted
COLT
2010
Springer
14 years 7 months ago
Best Arm Identification in Multi-Armed Bandits
We consider the problem of finding the best arm in a stochastic multi-armed bandit game. The regret of a forecaster is here defined by the gap between the mean reward of the optim...
Jean-Yves Audibert, Sébastien Bubeck, R&eac...