Sigdial 2019

Along with Dr. Zhou Yu and Arbit Chen (UCD Computer Science), we are thrilled that we have a paper accepted to the Special Interest Group on Discourse (SIGDIAL) meeting in Stockholm, Sweden.

Our paper explores how different text-to-speech (TTS) modifications to the 2018 Alexa Prize Winner chatbot, Gunrock, impact user ratings.

Cohn, M., Chen, C., Yu, Z. (2019). A Large-Scale User Study of an Alexa Prize Chatbot: Effect of TTS Dynamism on Perceived Quality of Social Dialog. (In press). 2019 Special Interest Group on Discourse and Dialogue, SIGDIAL. 

NSF Postdoctoral Fellowship (SPRF) Awardee

I am thrilled to serve as a PI for a two-year NSF-funded postdoctoral research fellowship with Drs. Georgia Zellou, Zhou Yu, and Katharine Graf Estes to explore human-voice AI interaction. (Click here to see the official NSF posting)

 

The project

We explore ways in which adults and children adapt their speech when talking to voice-activated digital assistants (e.g., Amazon’s Alexa), compared to adult human interlocutors.

This line of work provides a way to test differing theoretical predictions as to the extent that speech-register adjustments are driven by functional motives (e.g., intelligibility) and social factors (e.g., gender).

For instance, this research explores whether the same functional motivations that apply when correcting comprehension errors to human interlocutors apply in device-directed speech (DS), such as in manipulating the phonological nature of errors, to carefully control the level of intelligibility-related pressures in communication.

At the same time, this project explores how social factors may impact speech adaptation strategies, such as by interlocutor type, speaker age, or device gender. This project additionally involves important methodological innovations in programming and running experiments directly through a digital device platform.

Overall, this project aims to fill a gap in our knowledge in the acoustic-phonetic adjustments humans make when talking to voice-AI devices, and can ultimately reveal the underlying mechanisms in speech production by different speakers (e.g., based on age, gender, device experience), contributing to basic science research.

Interspeech 2019

We are excited that several papers have been accepted for the Interspeech 2019 meeting in Graz, Austria!

Papers on human-voice AI interaction

Cohn, M., & Zellou, G.(2019). Expressiveness influences human vocal alignment toward voice-AI. (In press). 2019 Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH. 

Snyder, C. Cohn, M., & Zellou, G. (2019). Individual variation in cognitive processing style predicts differences in phonetic imitation of device and human voices. (In press). 2019 Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH. 

Ferenc Segedin, C. Cohn, M., & Zellou, G. (2019). Perceptual adaptation to device and human voices:  learning and generalization of a phonetic shift across real and voice-AI talkers. (In press). 2019 Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH. 

Paper on musical training & speech perception

Cohn, M., Zellou, G., Barreda, S. (2019) The role of musical experience in the perceptual weighting of acoustic cues for the obstruent coda voicing contrast in American English. (In press). 2019 Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH. 

Amazon VUI Summit

This June, Dr. Georgia Zellou and I gave two invited talks at the June 2019 Voice User Interface (VUI) Summit at the Amazon headquarters. 

  • Zellou, G. Exploring human “speech rules” during vocal interactions with voice-AI. 
  • Cohn, M. Exploring cognitive-emotional expression: The impact of voice “emojis” in human-Alexa interaction.
Picture
Michelle Cohn (left) and Georgia Zellou (right) at the Amazon headquarters

Picture
Dr. Georgia Zellou

UCD Language Symposium

‘Most Innovative Research’ Panel

Undergraduate researcher, Melina Sarian, did a fantastic job presenting her research project at the ‘Most Innovative Research’ Panel. Her work extends our project exploring device expressiveness to human voices.

Sarian, M., Cohn, M., & Zellou, G. Human vocal alignment to voice­AI is mediated by acoustic expressiveness. [Talk].UC Davis Symposium on Language Research. Davis, CA.

Picture
Melina Sarian presenting at the ‘Most Innovative Research’ Panel

Picture
Melina Sarian presenting at the ‘Most Innovative Research’ Panel

Dynamics of Voice-AI Interaction Panel

Bruno Ferenc Segedin and I also presented two talks in our ‘Dynamics of Voice-AI Interaction’ ​panel

Cohn, M., Ferenc Segedin, B., & Zellou, G. Differences in cross­-generational prosodic alignment toward device and human voices [Talk]. UC Davis Symposium on Language Research. Davis, CA. 

Ferenc Segedin, B., Cohn, M., & Zellou, G. Perceptual adaptation to Amazon’s Alexa and human voices: asymmetries in learning and generalization of a novel accent across real and AI talkers. [Talk]. UC Davis Symposium on Language Research. Davis, CA. 

ICPhS 2019

We are thrilled that two papers from our lab were accepted for the 2019 International Congress of Phonetic Sciences meeting in Melbourne, Australia!

Cohn, M., Ferenc Segedin, B., Zellou, G. Imitating Siri: Socially-mediated vocal alignment to device and human voices. (In press). 2019 International Congress of Phonetic Sciences (ICPhS). 
Brotherton, C., Cohn, M., Zellou, G., Barreda, S. Sub-regional variation in positioning and degree of nasalization of /æ/ allophones in California (In press). 2019 International Congress of Phonetic Sciences (ICPhS).