Sciweavers

74 search results - page 1 / 15
» Regret Bounds for Gaussian Process Bandit Problems
Sort
View
JMLR
2010
125views more  JMLR 2010»
12 years 11 months ago
Regret Bounds for Gaussian Process Bandit Problems
Bandit algorithms are concerned with trading exploration with exploitation where a number of options are available but we can only learn their quality by experimenting with them. ...
Steffen Grünewälder, Jean-Yves Audibert,...
ICML
2010
IEEE
13 years 5 months ago
Gaussian Process Optimization in the Bandit Setting: No Regret and Experimental Design
Many applications require optimizing an unknown, noisy function that is expensive to evaluate. We formalize this task as a multiarmed bandit problem, where the payoff function is ...
Niranjan Srinivas, Andreas Krause, Sham Kakade, Ma...
CORR
2010
Springer
174views Education» more  CORR 2010»
13 years 4 months ago
Gaussian Process Bandits for Tree Search
We motivate and analyse a new Tree Search algorithm, based on recent advances in the use of Gaussian Processes for bandit problems. We assume that the function to maximise on the ...
Louis Dorard, John Shawe-Taylor
CORR
2011
Springer
202views Education» more  CORR 2011»
12 years 11 months ago
Online Least Squares Estimation with Self-Normalized Processes: An Application to Bandit Problems
The analysis of online least squares estimation is at the heart of many stochastic sequential decision-making problems. We employ tools from the self-normalized processes to provi...
Yasin Abbasi-Yadkori, Dávid Pál, Csa...
ALT
2009
Springer
14 years 1 months ago
Pure Exploration in Multi-armed Bandits Problems
Abstract. We consider the framework of stochastic multi-armed bandit problems and study the possibilities and limitations of strategies that explore sequentially the arms. The stra...
Sébastien Bubeck, Rémi Munos, Gilles...