Connect with us

Hi, what are you looking for?

Tech & Science

Are we falling out of love with voice assistants?

Users express dismay when interacting with smartphones, where frustration often arises due to misunderstandings between humans and their devices.

Amazon Echo speaker configured for the Alexa voice assistant. Image © Tim Sandle.
Amazon Echo speaker configured for the Alexa voice assistant. Image © Tim Sandle.

Voice assistants have stalled a little in terms of their development arc, with consumers commonly expecting more sophisticated algorithms. For example, 41 percent of voice assistant users have reported experiencing miscommunication issues with their devices every week.

This represents part of a data set compiled by the technology firm Cubicfox.

In a world becoming more reliant on voice recognition, clarity is crucial. Thomas M. Pentz, CEO of Cubicfox explains to Digital Journal: “The future of communication lies not just in the technology, but in our ability to understand and adapt to it.”

In particular, users express dismay when interacting with smartphones, where frustration often arises due to misunderstandings between humans and their devices.

Understanding the Gap

According to a report from Microsoft, smartphone users are increasingly frustrated with their devices’ ability to comprehend their voice commands. 41 percent of users express worries regarding trust, privacy, and passive listening. This underscores a discrepancy between user expectations and the present capabilities of speech recognition technology.

Reasons for misunderstandings

The reason  why a voice assistant sometimes fails to deliver an intelligible response include:

Background noise: Software algorithms responsible for voice recognition rely on clear audio input. When background noise creeps in, these algorithms face difficulty separating your voice from the surrounding sounds.

Accents and dialects: According to Frontier voice assistants like Alexa achieve a 55 percent accuracy rate for native speakers. However, this drops significantly for non-native speakers or those with strong accents.

Ambiguous phrasing: The way we naturally speak often involves incomplete sentences, slang, or informal language. These elements can be challenging for AI models to interpret compared to grammatically correct and formal phrasing.

A path forwards is with improved noise cancellation. It is likely that newer speech recognition models will integrate algorithms designed to effectively filter out background noise. Furthermore, developers are increasingly focusing on training models with diverse datasets that include various accents and dialects. This can lead to a more inclusive and accurate understanding of spoken language.

Avatar photo
Written By

Dr. Tim Sandle is Digital Journal's Editor-at-Large for science news. Tim specializes in science, technology, environmental, business, and health journalism. He is additionally a practising microbiologist; and an author. He is also interested in history, politics and current affairs.

You may also like:

News

The fountains of sleaze are overflowing. The pipelines can’t handle it.

Social Media

Meta's lawyers fought for six days in court to remove jurors who they deemed overly hostile to Facebook and Instagram.

Business

The Washington Post said Saturday its CEO and publisher Will Lewis was leaving effective immediately.

Business

US tariffs and shifting trade routes are meanwhile disrupting flows through key cutting and polishing hubs.