z-logo
open-access-imgOpen Access
Contribution of Prosody in Audio-Visual Integration to Emotional Perception of Virtual Characters
Author(s) -
Ekaterina Volkova,
Betty J. Mohler,
Sally A. Linkenauger,
Ivelina V. Alexandrova,
HH Bülthoff
Publication year - 2011
Publication title -
i-perception
Language(s) - English
Resource type - Journals
ISSN - 2041-6695
DOI - 10.1068/ic774
Subject(s) - facial expression , identification (biology) , modalities , prosody , emotional expression , psychology , natural (archaeology) , emotional prosody , emotion perception , perception , expression (computer science) , dynamics (music) , speech recognition , voice analysis , affective computing , cognitive psychology , computer science , motion (physics) , communication , human–computer interaction , artificial intelligence , history , social science , pedagogy , botany , archaeology , neuroscience , sociology , biology , programming language
Recent technology provides us with realistic looking virtual characters. Motion capture and elaborate mathematical models supply data for natural looking, controllable facial and bodily animations. With the help of computational linguistics and artificial intelligence, we can automatically assign emotional categories to appropriate stretches of text for a simulation of those social scenarios where verbal communication is important. All this makes virtual characters a valuable tool for creation of versatile stimuli for research on the integration of emotion information from different modalities. We conducted an audio-visual experiment to investigate the differential contributions of emotional speech and facial expressions on emotion identification. We used recorded and synthesized speech as well as dynamic virtual faces, all enhanced for seven emotional categories. The participants were asked to recognize the prevalent emotion of paired faces and audio. Results showed that when the voice was recorded, the vocalized emotion influenced participants' emotion identification more than the facial expression. However, when the voice was synthesized, facial expression influenced participants' emotion identification more than vocalized emotion. Additionally, individuals did worse on identifying either the facial expression or vocalized emotion when the voice was synthesized. Our experimental method can help to determine how to improve synthesized emotional speech

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom