Speech Prosody 2010
Chicago, IL, USA
Visual cues to speech prosody are available from a speaker's face; however the form and/or location of such cues are likely to be inconsistent across speakers. Given this, the question arises as to whether such cues are general enough to signal the same prosody information across speakers, and if so, where and what these cues are. To investigate this, this study used visual-visual and auditory-visual matching tasks requiring participants to select pairs of stimuli that were produced with the same prosody within- and across-speakers when visual information was limited to the upper or lower face. Experiment 1 tested within-speaker prosody matching when the speaker's lower face was presented. The results showed highly accurate matching performance. Taken together with the results of our previous study which presented the upper face in the same tasks , these data provided a baseline for which to evaluate cross-speaker prosody matching (Experiment 2). In Experiment 2, both lower and upper face stimuli were presented. In comparison to within-speaker matching, performance was lower for cross-speaker matching but still greater than chance. Overall, the results suggest that both the upper and lower face provide general non-speaker specific as well as speaker-specific visual cues to prosody.
Index Terms: visual prosody, perception, cross-speaker, within-speaker, face area, narrow focus, echoic questions
Bibliographic reference. Cvejic, Erin / Kim, Jeesun / Davis, Chris (2010): "It's all the same to me: prosodic discrimination across speakers and face areas", In SP-2010, paper 893.