

Clear speech in the new digital era: Speaking and listening clearly to voice-AI...
source link: https://techxplore.com/news/2023-05-speech-digital-era-voice-ai.html
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.

Clear speech in the new digital era: Speaking and listening clearly to voice-AI systems
by Acoustical Society of America

Millions of people now regularly communicate with AI-based devices, such as smartphones, speakers, and cars. Studying these interactions can improve AI's ability to understand human speech and determine how talking with technology impacts language.
In their talk, "Clear speech in the new digital era: Speaking and listening clearly to voice-AI systems," Georgia Zellou and Michelle Cohn of the University of California, Davis described experiments to investigate how speech and comprehension change when humans communicate with AI. The presentation took place as part of the 184th Meeting of the Acoustical Society of America running May 8-12.
In their first line of questioning, Zellou and Cohn examined how people adjust their voice when communicating with an AI system compared to talking with another human. They found the participants produced louder and slower speech with less pitch variation when they spoke to voice-AI (e.g., Siri, Alexa), even across identical interactions.
On the listening side, the researchers showed that how humanlike a device sounds impacts how well listeners will understand it. If a listener thinks the voice talking is a device, they are less able to accurately understand. However, if it sounds more humanlike, their comprehension increases. Clear speech, like in the style of a newscaster, was better understood overall, even if it was machine-generated.
"We do see some differences in patterns across human- and machine-directed speech: People are louder and slower when talking to technology. These adjustments are similar to the changes speakers make when talking in background noise, such as in a crowded restaurant," said Zellou. "People also have expectations that the systems will misunderstand them and that they won't be able to understand the output."
Clarifying what makes a speaker intelligible will be useful for voice technology. For example, these results suggest that text-to-speech voices should adopt a "clear" style in noisy conditions.
Looking forward, the team aims to apply these studies to people from different age groups and social and language backgrounds. They also want to investigate how people learn language from devices and how linguistic behavior adapts as technology changes.
"There are so many open questions," said Cohn. "For example, could voice-AI be a source of language change among some speakers? As technology advances, such as with large language models like ChatGPT, the boundary between human and machine is changing—how will our language change with it?"
More information: Conference: acousticalsociety.org/asa-meetings/
Recommend
-
32
Android With Sound Amplifier, more people can hear clearly ...
-
38
Audio samples from "Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis" Paper: arXiv...
-
7
Speaking at C# Corner’s Virtual Conference C# Corner is holding
-
4
Shhhh, They’re Listening – Inside the Coming Voice-Profiling RevolutionYves here. This article confirms my prejudices about the importance of avoiding those spying home assistants at all costs. And it takes a bit of effort to try to thwart fi...
-
11
Rejected for a Speaking Submission I got a speaking rejection today. Well, not today, but the day I wrote this. Not...
-
4
Evidence of a predictive coding hierarchy in the human brain listening to speechAbstractConsiderable progress has recently been made in natural language proce...
-
14
Support is great. Feedback is even better."Thanks for checking us out - We would love to hear from you. Please share your feedback and suggestions. We are working very hard to make the product better :)"The makers of Au...
-
17
Gliglish · Your AI Language Teacher
-
3
Another NAS-ty WD controversy — “Clearly predatory”: Western Digital sparks panic, anger for age-shaming HDDs Drives automatically get a "warning" flag if powered on for 3...
-
5
Media In social media, speaking without listening is malpractice ByJosh Bernof...
About Joyk
Aggregate valuable and interesting links.
Joyk means Joy of geeK