NUK - logo
E-resources
Peer reviewed Open access
  • Prosodic alignment toward e...
    Cohn, Michelle; Predeck, Kristin; Sarian, Melina; Zellou, Georgia

    Speech communication, December 2021, 2021-12-00, 20211201, Volume: 135
    Journal Article

    •Participants shadow emotionally expressive and neutral productions•Comparison of speech toward human and Amazon Alexa model talkers•Evidence of alignment toward emotional style for duration, mean f0, and f0 variation•Acoustic differences across model stimuli predict alignment patterns•Largely similar patterns by female and male participants This study tests whether individuals vocally align toward emotionally expressive prosody produced by two types of interlocutors: a human and a voice-activated artificially intelligent (voice-AI) assistant. Participants completed a word shadowing experiment of interjections (e.g., “Awesome”) produced in emotionally neutral and expressive prosodies by both a human voice and a voice generated by a voice-AI system (Amazon's Alexa). Results show increases in participants’ word duration, mean f0, and f0 variation in response to emotional expressiveness, consistent with increased alignment toward a general ‘positive-emotional’ speech style. Small differences in emotional alignment by talker category (human vs. voice-AI) parallel the acoustic differences in the model talkers’ productions, suggesting that participants are mirroring the acoustics they hear. The similar responses to emotion in both a human and voice-AI talker support accounts of unmediated emotional alignment, as well as computer personification: people apply emotionally-mediated behaviors to both types of interlocutors. While there were small differences in magnitude by participant gender, the overall patterns were similar for women and men, supporting a nuanced picture of emotional vocal alignment.