Sciweavers

Share
MICAI
2010
Springer

Combining Neural Networks Based on Dempster-Shafer Theory for Classifying Data with Imperfect Labels

8 years 2 months ago
Combining Neural Networks Based on Dempster-Shafer Theory for Classifying Data with Imperfect Labels
This paper addresses the supervised learning in which the class membership of training data are subject to uncertainty. This problem is tackled in the framework of the Dempster-Shafer theory. In order to properly estimate the class labels, different types of features are extracted from the data. The initial labels of the training data are ignored and by utilizing the main classes' prototypes, each training pattern, in each of the feature spaces, is reassigned to one class or a subset of the main classes based on the level of ambiguity concerning its class label. Multilayer perceptrons neural network is used as base classifier and for a given test sample, its outputs are considered as basic belief assignment. Finally, the decisions of the base classifiers are combined using Dempster's rule of combination. Experiments with artificial and real data demonstrate that considering ambiguity in class labels can provide better results than classifiers trained with imperfect labels.
Mahdi Tabassian, Reza Ghaderi, Reza Ebrahimpour
Added 29 Jan 2011
Updated 29 Jan 2011
Type Journal
Year 2010
Where MICAI
Authors Mahdi Tabassian, Reza Ghaderi, Reza Ebrahimpour
Comments (0)
books