Visual Contribution to Speech Perception: Measuring the Intelligibility of Animated Talking Heads
Author(s) -
Slim Ouni,
Michael M. Cohen,
Hope Ishak,
Dominic W. Massaro
Publication year - 2006
Publication title -
eurasip journal on audio speech and music processing
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.259
H-Index - 29
eISSN - 1687-4722
pISSN - 1687-4714
DOI - 10.1155/2007/47891
Subject(s) - intelligibility (philosophy) , computer science , perception , natural (archaeology) , speech recognition , natural language processing , artificial intelligence , psychology , history , philosophy , epistemology , neuroscience , archaeology
Animated agents are becoming increasingly frequent in research and applications in speech science. An important challenge is to evaluate the effectiveness of the agent in terms of the intelligibility of its visible speech. In three experiments, we extend and test the Sumby and Pollack (1954) metric to allow the comparison of an agent relative to a standard or reference, and also propose a new metric based on the fuzzy logical model of perception (FLMP) to describe the benefit provided by a synthetic animated face relative to the benefit provided by a natural face. A valid metric would allow direct comparisons accross different experiments and wouldgive measures of the benfit of a synthetic animated face relative to a natural face (or indeed any two conditions) and how this benefit varies as a function of the type of synthetic face, the test items (e.g., syllables versus sentences), different individuals, and applications
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom