Sciweavers

MLMI
2004
Springer

Mapping from Speech to Images Using Continuous State Space Models

13 years 9 months ago
Mapping from Speech to Images Using Continuous State Space Models
In this paper a system that transforms speech waveforms to animated faces are proposed. The system relies on continuous state space models to perform the mapping, this makes it possible to ensure video with no sudden jumps and allows continuous control of the parameters in ’face space’. The performance of the system is critically dependent on the number of hidden variables, with too few variables the model cannot represent data, and with too many overfitting is noticed. Simulations are performed on recordings of 3-5 sec. video sequences with sentences from the Timit database. From a subjective point of view the model is able to construct an image sequence from an unknown noisy speech sequence even though the number of training examples are limited.
Tue Lehn-Schiøler, Lars Kai Hansen, Jan Lar
Added 02 Jul 2010
Updated 02 Jul 2010
Type Conference
Year 2004
Where MLMI
Authors Tue Lehn-Schiøler, Lars Kai Hansen, Jan Larsen
Comments (0)