WinkTalk: a demonstration of a multimodal speech synthesis platform linking facial expressions to expressive synthetic voices
2012 (English)In: Proceedings of the Third Workshop on Speech and Language Processing for Assistive Technologies, Association for Computational Linguistics , 2012, 5-8 p.Conference paper (Refereed)Text
This paper describes a demonstration of the WinkTalk system, which is a speech synthe- sis platform using expressive synthetic voices. With the help of a webcamera and facial ex- pression analysis, the system allows the user to control the expressive features of the syn- thetic speech for a particular utterance with their facial expressions. Based on a person- alised mapping between three expressive syn- thetic voices and the users facial expressions, the system selects a voice that matches their face at the moment of sending a message. The WinkTalk system is an early research pro- totype that aims to demonstrate that facial expressions can be used as a more intuitive control over expressive speech synthesis than manual selection of voice types, thereby con- tributing to an improved communication expe- rience for users of speech generating devices.
Place, publisher, year, edition, pages
Association for Computational Linguistics , 2012. 5-8 p.
Engineering and Technology
IdentifiersURN: urn:nbn:se:kth:diva-185524OAI: oai:DiVA.org:kth-185524DiVA: diva2:922775
Third Workshop on Speech and Language Processing for Assistive Technologies
QC 201604252016-04-252016-04-212016-04-25Bibliographically approved