EmoShapelets: Capturing local dynamics of audio-visual affective speech

4 years 9 months ago
EmoShapelets: Capturing local dynamics of audio-visual affective speech
—Automatic recognition of emotion in speech is an active area of research. One of the important open challenges relates to how the emotional characteristics of speech change in time. Past research has demonstrated the importance of capturing global dynamics (across an entire utterance) and local dynamics (within segments of an utterance). In this paper, we propose a novel concept, EmoShapelets, to capture the local dynamics in speech. EmoShapelets capture changes in emotion that occur within utterances. We propose a framework to generate, update, and select EmoShapelets. We also demonstrate the discriminative power of EmoShapelets by using them with various classifiers to achieve comparable results with the state-of-the-art systems on the IEMOCAP dataset. EmoShapelets can serve as basic units of emotion expression and provide additional evidence supporting the existence of local patterns of emotion underlying human communication. Keywords—emotion classification; time series; emot...
Added 13 Apr 2016
Updated 13 Apr 2016
Type Journal
Year 2015
Where ACII
Comments (0)