Through all three of Professor Sumner’s papers and the worksheet
we looked at in class, there is one overarching theme: phonetic variation. We
all know that speech can gives our brains information about sounds and words. These
papers provide proof that speech also feeds our brains information about the
speaker – we make inferences about their age, gender, accent, emotion, and
style. We take in sensory information automatically and draw conclusions about
speakers even when we do not mean to. This is due to social weighting. Different
accents can change meaning for people. We may think someone with one accent is
telling a lie, or we may only get the gist of what someone is saying, or not
recognize specific words when everything is opposite if it is said by someone
with a different accent. Responses to something someone with one accent may say
can be more variable if someone with a different accent said it.
All this information makes me question what, then, would be
best for Siri to use? Should she try to mimic the speaker’s accent? Would it be
a safe bet to have her always use General American? Research shows that GA is
most easily digestible for most people, even those with accents. But would
people believe everything she is saying if the listener speaks a different
accent?
I also know people who have Siri speak with a British
accent. Prestigious accents usually allow people to remember exact phrasing.
Does this suggest that perhaps British is the best option for Siri? This would
allow us to better remember her responses which would help us with directions
and other information we asked for.
With different groups of speakers responding differently to
various accents, perhaps there is no “best” accent for AI to use. However, it
would be interesting to see if technology using voices works better or incites
quicker responses if it uses the same accent as its individual user.
Hey Andrea,
ReplyDeleteYour comment on the practical application of Sumner's research on the voice-recognition technology we used today was very interesting! What I'd be curious about the most is how much it matters to have the "optimal" accent for speaking. Does it make that much of a marked difference than the American (or Australian) accent? Also, I think it'd be interesting to examine perhaps the novelty of a voice and how it impacts our encoding and behaviors. In particular, I am talking about the navigation app Waze, which allows your turn-by-turn instructions to be narrated by a huge list of celebrities, including Morgan Freeman, Arnold Schwarzenegger, and Stephen Colbert.Sure, it is fun to have a celebrity seemingly navigate you, but perhaps there is some science in how listening to a distinct voice (different than the normal robotic voice that we typically hear) can improve our abilities to encode directions into our minds and consquently and not look at our phones while driving? I tried to do some Googling to see if any research has been done on this and it doesn't look like there has been. Imagine the positive PR this could bring if it actually was shown that having these wacky celebrity voices actually improves the safety of viewers!