Audio-visual synthesized attitudes presented by the German speaking robot SMiRAE
Hönemann A, Bennett C, Wagner P, Sabanovic S (2019)
In: Proceedings of the 15th International Conference on Auditory-Visual Speech Processing.
Konferenzbeitrag | Englisch
Download
Es wurden keine Dateien hochgeladen. Nur Publikationsnachweis!
Autor*in
Hönemann, Angelika;
Bennett, Casey;
Wagner, PetraUniBi ;
Sabanovic, Selma
Einrichtung
Abstract / Bemerkung
This paper presents the acoustic and visual modeling of nine attitudinal expressions that were realized by the German speaking robot SMiRAE which is a speech-enabled version of the non-speaking robotic face MiRAE previously developed at Indiana University. The parameter-oriented acoustic model is based on the German Mary TTS which is part of the speech processing system InproTK. Visual realization of expressions is based on five defined basic emotions of the Facial Action Coding System (FACS) developed by Ekman. Both models were additionally modified with respect to results of an audio-visual analysis and evaluation of human portrayals of attitudes recorded in our previous work.
The plausibility of synthesized attitudinal expressions is shown by an association study in which 18 participants described 54 attitudes in a free association. Basis for a 5-cluster classification was the first four dimensions of a correspondence analysis which accounted 78% of variance in participant perception. Significant correlations were seen between 66 normalized participant descriptions and the robot’s displayed attitudes. For instance, the attitudes admiration and politeness were associated with the terms freundlich and gluecklich, the interrogative attitudes surprise and doubt with the terms fragend, verwundert and skeptisch, the expression uncertatinty was perceived with traurig and besorgt.
Stichworte
speech production and perception;
attitudinal speech;
acoustic and visual synthesis;
human-robot interaction;
biphonetics
Erscheinungsjahr
2019
Titel des Konferenzbandes
Proceedings of the 15th International Conference on Auditory-Visual Speech Processing
Konferenz
15th International Conference on Auditory-Visual Speech Processing
Konferenzort
Melbourne, Australia
Konferenzdatum
2019-09-10 – 2019-08-11
Page URI
https://pub.uni-bielefeld.de/record/2936383
Zitieren
Hönemann A, Bennett C, Wagner P, Sabanovic S. Audio-visual synthesized attitudes presented by the German speaking robot SMiRAE. In: Proceedings of the 15th International Conference on Auditory-Visual Speech Processing. 2019.
Hönemann, A., Bennett, C., Wagner, P., & Sabanovic, S. (2019). Audio-visual synthesized attitudes presented by the German speaking robot SMiRAE. Proceedings of the 15th International Conference on Auditory-Visual Speech Processing
Hönemann, Angelika, Bennett, Casey, Wagner, Petra, and Sabanovic, Selma. 2019. “Audio-visual synthesized attitudes presented by the German speaking robot SMiRAE”. In Proceedings of the 15th International Conference on Auditory-Visual Speech Processing.
Hönemann, A., Bennett, C., Wagner, P., and Sabanovic, S. (2019). “Audio-visual synthesized attitudes presented by the German speaking robot SMiRAE” in Proceedings of the 15th International Conference on Auditory-Visual Speech Processing.
Hönemann, A., et al., 2019. Audio-visual synthesized attitudes presented by the German speaking robot SMiRAE. In Proceedings of the 15th International Conference on Auditory-Visual Speech Processing.
A. Hönemann, et al., “Audio-visual synthesized attitudes presented by the German speaking robot SMiRAE”, Proceedings of the 15th International Conference on Auditory-Visual Speech Processing, 2019.
Hönemann, A., Bennett, C., Wagner, P., Sabanovic, S.: Audio-visual synthesized attitudes presented by the German speaking robot SMiRAE. Proceedings of the 15th International Conference on Auditory-Visual Speech Processing. (2019).
Hönemann, Angelika, Bennett, Casey, Wagner, Petra, and Sabanovic, Selma. “Audio-visual synthesized attitudes presented by the German speaking robot SMiRAE”. Proceedings of the 15th International Conference on Auditory-Visual Speech Processing. 2019.