Sciweavers

TMM
2011

Audiovisual Discrimination Between Speech and Laughter: Why and When Visual Information Might Help

12 years 11 months ago
Audiovisual Discrimination Between Speech and Laughter: Why and When Visual Information Might Help
Past research on automatic laughter classification / detection has focused mainly on audio-based approaches. Here we present an audiovisual approach to distinguishing laughter from speech and we show that integrating the information from audio and video channels may lead to improved performance over single-modal approaches. Both audio and visual channels consist of two streams (cues), facial expressions and head pose for video, and cepstral and prosodic features for audio. Two types of experiments were performed: 1) subject-independent cross-validation on the AMI dataset, and 2) cross-database experiments on the AMI and SAL datasets. We experimented with different combinations of cues with the most informative being the combination of facial expressions, cepstral and prosodic features. Our results suggest that the performance of the audiovisual approach is better on average than single-modal approaches. The addition of visual information produces better results when it comes to femal...
Stavros Petridis, Maja Pantic
Added 15 May 2011
Updated 15 May 2011
Type Journal
Year 2011
Where TMM
Authors Stavros Petridis, Maja Pantic
Comments (0)