Sciweavers

Share
ICMLA
2009

Learning Deep Neural Networks for High Dimensional Output Problems

9 years 8 months ago
Learning Deep Neural Networks for High Dimensional Output Problems
State-of-the-art pattern recognition methods have difficulty dealing with problems where the dimension of the output space is large. In this article, we propose a new framework based on deep architectures (e.g. Deep Neural Networks) in order to deal with this issue. Deep architectures have proven to be efficient for high dimensional input problems such as image classification, due to their ability to embed the input space. The main contribution of this article is the extension of the embedding procedure to both the input and output spaces in order to easily handle high dimensional output problems. Using this extension, interoutput dependencies can be modelled efficiently. This provides an interesting alternative to probabilistic models such as HMM and CRF. Preliminary experiments on toy datasets and USPS character reconstruction show promising results.
Benjamin Labbé, Romain Hérault, Cl&e
Added 19 Feb 2011
Updated 19 Feb 2011
Type Journal
Year 2009
Where ICMLA
Authors Benjamin Labbé, Romain Hérault, Clément Chatelain
Comments (0)
books