Sciweavers

Share
IJCNN
2006
IEEE

Nonlinear principal component analysis of noisy data

3 years 10 months ago
Nonlinear principal component analysis of noisy data
With very noisy data, having plentiful samples eliminates overfitting in nonlinear regression, but not in nonlinear principal component analysis (NLPCA). To overcome this problem in NLPCA, a new information criterion (IC) is proposed for selecting the best model among multiple models with different complexity and regularization (i.e. weight penalty). This IC gauges the inconsistency I between the nonlinear principal components (u and ˜u) for every data point x and its nearest neighbour ˜x, with I = 1 − correlation(u, ˜u), where I tends to increase with overfitted solutions. Tests were performed using autoassociative neural networks for NLPCA on synthetic and real climate data (tropical Pacific sea surface temperatures and equatorial stratospheric winds), with the IC performing well in model selection and in deciding between an open curve or a closed curve solution.
William W. Hsieh
Added 11 Jun 2010
Updated 11 Jun 2010
Type Conference
Year 2006
Where IJCNN
Authors William W. Hsieh
Comments (0)
books