Sciweavers

6 search results - page 1 / 2
» A visually grounded natural language interface for reference...
Sort
View
ICMI
2003
Springer
164views Biometrics» more  ICMI 2003»
13 years 9 months ago
A visually grounded natural language interface for reference to spatial scenes
Many user interfaces, from graphic design programs to navigation aids in cars, share a virtual space with the user. Such applications are often ideal candidates for speech interfa...
Peter Gorniak, Deb Roy
CSL
2002
Springer
13 years 4 months ago
Learning visually grounded words and syntax for a scene description task
A spoken language generation system has been developed that learns to describe objects in computer-generated visual scenes. The system is trained by a `show-and-tell' procedu...
Deb K. Roy
ICMI
2010
Springer
217views Biometrics» more  ICMI 2010»
13 years 2 months ago
Focusing computational visual attention in multi-modal human-robot interaction
Identifying verbally and non-verbally referred-to objects is an important aspect of human-robot interaction. Most importantly, it is essential to achieve a joint focus of attentio...
Boris Schauerte, Gernot A. Fink
AVI
2006
13 years 6 months ago
Oral messages improve visual search
Input multimodality combining speech and hand gestures has motivated numerous usability studies. Contrastingly, issues relating to the design and ergonomic evaluation of multimoda...
Suzanne Kieffer, Noelle Carbonell
HRI
2006
ACM
13 years 10 months ago
Structural descriptions in human-assisted robot visual learning
The paper presents an approach to using structural descriptions, obtained through a human-robot tutoring dialogue, as labels for the visual object models a robot learns. The paper...
Geert-Jan M. Kruijff, John D. Kelleher, Gregor Ber...