Sciweavers

5 search results - page 1 / 1
» Finite-Sample Convergence Rates for Q-Learning and Indirect ...
Sort
View
NIPS
1998
13 years 6 months ago
Finite-Sample Convergence Rates for Q-Learning and Indirect Algorithms
In this paper, we address two issues of long-standing interest in the reinforcement learning literature. First, what kinds of performance guarantees can be made for Q-learning aft...
Michael J. Kearns, Satinder P. Singh
NIPS
2007
13 years 6 months ago
Cluster Stability for Finite Samples
Over the past few years, the notion of stability in data clustering has received growing attention as a cluster validation criterion in a sample-based framework. However, recent w...
Ohad Shamir, Naftali Tishby
TIT
2002
86views more  TIT 2002»
13 years 4 months ago
Lagrangian empirical design of variable-rate vector quantizers: consistency and convergence rates
Abstract--The Lagrangian formulation of variable-rate vector quantization is known to yield useful necessary conditions for quantizer optimality and generalized Lloyd algorithms fo...
Tamás Linder
NIPS
2008
13 years 6 months ago
Unlabeled data: Now it helps, now it doesn't
Empirical evidence shows that in favorable situations semi-supervised learning (SSL) algorithms can capitalize on the abundance of unlabeled training data to improve the performan...
Aarti Singh, Robert D. Nowak, Xiaojin Zhu
CVPR
2008
IEEE
14 years 6 months ago
Generalised blurring mean-shift algorithms for nonparametric clustering
Gaussian blurring mean-shift (GBMS) is a nonparametric clustering algorithm, having a single bandwidth parameter that controls the number of clusters. The algorithm iteratively sh...
Miguel Á. Carreira-Perpiñán