We introduce a novel approach to modeling the dynamics of human facial motion induced by the action of speech for the purpose of synthesis. We represent the trajectories of a numbe...
In this paper we introduce a system that automatically adds different types of non-verbal behavior to a given dialogue script between two virtual embodied agents. It allows us to t...
Werner Breitfuss, Helmut Prendinger, Mitsuru Ishiz...
An essential step in the generation of expressive speech synthesis is the automatic detection and classification of emotions most likely to be present in textual input. At last I...
We describe a method for the fully automatic learning of hierarchical finite state translation models. The input to the method is transcribed speech utterances and their correspon...
We present MikeTalk, a text-to-audiovisual speech synthesizer which converts input text into an audiovisual speech stream. MikeTalk is built using visemes, which are a set of imag...