Sciweavers

Share
ECAI
2000
Springer

Learning Efficiently with Neural Networks: A Theoretical Comparison between Structured and Flat Representations

11 years 9 months ago
Learning Efficiently with Neural Networks: A Theoretical Comparison between Structured and Flat Representations
Abstract. We are interested in the relationship between learning efficiency and representation in the case of supervised neural networks for pattern classification trained by continuous error minimization techniques, such as gradient descent. In particular, we focus our attention on a recently introduced architecture called recursive neural network (RNN) which is able to learn class membership of patterns represented as labeled directed ordered acyclic graphs (DOAG). RNNs offer several benefits compared to feedforward and recurrent networks for sequences. However, how RNNs compare to these models in terms of learning efficiency still needs investigation. In this paper we give a theoretical answer by giving a set of results concerning the shape of the error surface and critically discussing the implications of these results on the relative difficulty of learning with different data representations. The message of this paper is that, whenever structured representations are available, the...
Marco Gori, Paolo Frasconi, Alessandro Sperduti
Added 24 Aug 2010
Updated 24 Aug 2010
Type Conference
Year 2000
Where ECAI
Authors Marco Gori, Paolo Frasconi, Alessandro Sperduti
Comments (0)
books