Sciweavers

17 search results - page 2 / 4
» The Budgeted Multi-armed Bandit Problem
Sort
View
CORR
2006
Springer
83views Education» more  CORR 2006»
13 years 5 months ago
How to Beat the Adaptive Multi-Armed Bandit
The multi-armed bandit is a concise model for the problem of iterated decision-making under uncertainty. In each round, a gambler must pull one of K arms of a slot machine, withou...
Varsha Dani, Thomas P. Hayes
CORR
2010
Springer
143views Education» more  CORR 2010»
13 years 1 months ago
The Non-Bayesian Restless Multi-Armed Bandit: a Case of Near-Logarithmic Regret
In the classic Bayesian restless multi-armed bandit (RMAB) problem, there are N arms, with rewards on all arms evolving at each time as Markov chains with known parameters. A play...
Wenhan Dai, Yi Gai, Bhaskar Krishnamachari, Qing Z...
CORR
2010
Springer
187views Education» more  CORR 2010»
13 years 5 months ago
Learning in A Changing World: Non-Bayesian Restless Multi-Armed Bandit
We consider the restless multi-armed bandit (RMAB) problem with unknown dynamics. In this problem, at each time, a player chooses K out of N (N > K) arms to play. The state of ...
Haoyang Liu, Keqin Liu, Qing Zhao
CORR
2010
Springer
152views Education» more  CORR 2010»
13 years 3 days ago
Combinatorial Network Optimization with Unknown Variables: Multi-Armed Bandits with Linear Rewards
In the classic multi-armed bandits problem, the goal is to have a policy for dynamically operating arms that each yield stochastic rewards with unknown means. The key metric of int...
Yi Gai, Bhaskar Krishnamachari, Rahul Jain
COLT
2010
Springer
13 years 3 months ago
Best Arm Identification in Multi-Armed Bandits
We consider the problem of finding the best arm in a stochastic multi-armed bandit game. The regret of a forecaster is here defined by the gap between the mean reward of the optim...
Jean-Yves Audibert, Sébastien Bubeck, R&eac...