When human listeners utter Listener Responses (e.g. back-channels or acknowledgments) such as ‘yeah’ and ‘mmhmm’, interlocutors commonly continue to speak or resume their ...
We introduce MySong, a system that automatically chooses chords to accompany a vocal melody. A user with no musical experience can create a song with instrumental accompaniment ju...
Automated analysis of human affective behavior has attracted increasing attention from researchers in psychology, computer science, linguistics, neuroscience, and related discipli...
Zhihong Zeng, Maja Pantic, Glenn I. Roisman, Thoma...
Studying the relationship between natural language and affective information as well as assessing the underpinned affective qualities of natural language are becoming crucial for ...
Shaikh Mostafa Al Masum, Helmut Prendinger, Mitsur...
Human tutors detect and respond to student emotional states, but current machine tutors do not. Our preliminary machine learning experiments involving transcription, emotion annot...