Sciweavers

TNN
1998

A class of competitive learning models which avoids neuron underutilization problem

13 years 4 months ago
A class of competitive learning models which avoids neuron underutilization problem
— In this paper, we study a qualitative property of a class of competitive learning (CL) models, which is called the multiplicatively biased competitive learning (MBCL) model, namely that it avoids neuron underutilization with probability one as time goes to infinity. In the MBCL, the competition among neurons is biased by a multiplicative term, while only one weight vector is updated per learning step. This is of practical interest since its instances have computational complexities among the lowest in existing CL models. In addition, in applications like classification, vector quantizer design and probability density function estimation, a necessary condition for optimal performance is to avoid neuron underutilization. Hence, it is possible to define instances of MBCL to achieve optimal performance in these applications.
Clifford Sze-Tsan Choy, Wan-Chi Siu
Added 23 Dec 2010
Updated 23 Dec 2010
Type Journal
Year 1998
Where TNN
Authors Clifford Sze-Tsan Choy, Wan-Chi Siu
Comments (0)