For applications with consecutive incoming training examples, on-line learning has the potential to achieve a likelihood as high as off-line learning without scanning all availabl...
A variation of the differential steepest descent algorithm, here called the dithered linear search (DLS), is examined and applied to analog filter adaptation. The DLS algorithm is...
Gradient-following learning methods can encounter problems of implementation in many applications, and stochastic variants are frequently used to overcome these difficulties. We ...
Stochastic gradient descent (SGD) uses approximate gradients estimated from subsets of the training data and updates the parameters in an online fashion. This learning framework i...
— Several recent algorithms have formulated the SLAM problem in terms of non-linear pose graph optimization. These algorithms are attractive because they offer lower computationa...