Change search
ReferencesLink to record
Permanent link

Direct link
Facial expression as an input annotation modality for affective speech-to-speech translation
2012 (English)Conference paper (Refereed)Text
Abstract [en]

One of the challenges of speech-to-speech translation is to accurately preserve the paralinguistic information in the speaker’s message. In this work we explore the use of automatic facial expression analysis as an input annotation modality to transfer paralinguistic information at a symbolic level from input to output in speech-to-speech translation. To evaluate the feasibility of this ap- proach, a prototype system, FEAST (Facial Expression-based Affective Speech Translation) has been developed. FEAST classifies the emotional state of the user and uses it to render the translated output in an appropriate voice style, using expressive speech synthesis. 

Place, publisher, year, edition, pages
National Category
Engineering and Technology
URN: urn:nbn:se:kth:diva-185527OAI: diva2:922771
Workshop on Multimodal Analyses enabling Artificial Agents in Human-Machine Interaction

QC 20160426

Available from: 2016-04-25 Created: 2016-04-21 Last updated: 2016-06-02Bibliographically approved

Open Access in DiVA

fulltext(853 kB)9 downloads
File information
File name FULLTEXT01.pdfFile size 853 kBChecksum SHA-512
Type fulltextMimetype application/pdf

Search in DiVA

By author/editor
Székely, Éva
Engineering and Technology

Search outside of DiVA

GoogleGoogle Scholar
Total: 9 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

Total: 6 hits
ReferencesLink to record
Permanent link

Direct link