Active Learning in the Non-realizable Case

12 years 6 months ago
Active Learning in the Non-realizable Case
Most of the existing active learning algorithms are based on the realizability assumption: The learner’s hypothesis class is assumed to contain a target function that perfectly classifies all training and test examples. This assumption can hardly ever be justified in practice. In this paper, we study how relaxing the realizability assumption affects the sample complexity of active learning. First, we extend existing results on query learning to show that any active learning algorithm for the realizable case can be transformed to tolerate random bounded rate class noise. Thus, bounded rate class noise adds little extra complications to active learning, and in particular exponential label complexity savings over passive learning are still possible. However, it is questionable whether this noise model is any more realistic in practice than assuming no noise at all. Our second result shows that if we move to the truly non-realizable model of statistical learning theory, then the label...
Matti Kääriäinen
Added 14 Mar 2010
Updated 14 Mar 2010
Type Conference
Year 2006
Where ALT
Authors Matti Kääriäinen
Comments (0)