Sciweavers

23 search results - page 3 / 5
» Online Optimization in X-Armed Bandits
Sort
View
NIPS
2004
13 years 6 months ago
Nearly Tight Bounds for the Continuum-Armed Bandit Problem
In the multi-armed bandit problem, an online algorithm must choose from a set of strategies in a sequence of n trials so as to minimize the total cost of the chosen strategies. Wh...
Robert D. Kleinberg
CORR
2010
Springer
127views Education» more  CORR 2010»
13 years 5 months ago
Online Algorithms for the Multi-Armed Bandit Problem with Markovian Rewards
We consider the classical multi-armed bandit problem with Markovian rewards. When played an arm changes its state in a Markovian fashion while it remains frozen when not played. Th...
Cem Tekin, Mingyan Liu
CORR
2010
Springer
171views Education» more  CORR 2010»
12 years 12 months ago
Online Learning in Opportunistic Spectrum Access: A Restless Bandit Approach
We consider an opportunistic spectrum access (OSA) problem where the time-varying condition of each channel (e.g., as a result of random fading or certain primary users' activ...
Cem Tekin, Mingyan Liu
COLT
2008
Springer
13 years 6 months ago
Regret Bounds for Sleeping Experts and Bandits
We study on-line decision problems where the set of actions that are available to the decision algorithm vary over time. With a few notable exceptions, such problems remained larg...
Robert D. Kleinberg, Alexandru Niculescu-Mizil, Yo...
LION
2010
Springer
190views Optimization» more  LION 2010»
13 years 9 months ago
Algorithm Selection as a Bandit Problem with Unbounded Losses
Abstract. Algorithm selection is typically based on models of algorithm performance learned during a separate offline training sequence, which can be prohibitively expensive. In r...
Matteo Gagliolo, Jürgen Schmidhuber