Connect with us

Hi, what are you looking for?

Tech & Science

Are we falling out of love with voice assistants?

Users express dismay when interacting with smartphones, where frustration often arises due to misunderstandings between humans and their devices.

Amazon Echo speaker configured for the Alexa voice assistant. Image © Tim Sandle.
Amazon Echo speaker configured for the Alexa voice assistant. Image © Tim Sandle.

Voice assistants have stalled a little in terms of their development arc, with consumers commonly expecting more sophisticated algorithms. For example, 41 percent of voice assistant users have reported experiencing miscommunication issues with their devices every week.

This represents part of a data set compiled by the technology firm Cubicfox.

In a world becoming more reliant on voice recognition, clarity is crucial. Thomas M. Pentz, CEO of Cubicfox explains to Digital Journal: “The future of communication lies not just in the technology, but in our ability to understand and adapt to it.”

In particular, users express dismay when interacting with smartphones, where frustration often arises due to misunderstandings between humans and their devices.

Understanding the Gap

According to a report from Microsoft, smartphone users are increasingly frustrated with their devices’ ability to comprehend their voice commands. 41 percent of users express worries regarding trust, privacy, and passive listening. This underscores a discrepancy between user expectations and the present capabilities of speech recognition technology.

Reasons for misunderstandings

The reason  why a voice assistant sometimes fails to deliver an intelligible response include:

Background noise: Software algorithms responsible for voice recognition rely on clear audio input. When background noise creeps in, these algorithms face difficulty separating your voice from the surrounding sounds.

Accents and dialects: According to Frontier voice assistants like Alexa achieve a 55 percent accuracy rate for native speakers. However, this drops significantly for non-native speakers or those with strong accents.

Ambiguous phrasing: The way we naturally speak often involves incomplete sentences, slang, or informal language. These elements can be challenging for AI models to interpret compared to grammatically correct and formal phrasing.

A path forwards is with improved noise cancellation. It is likely that newer speech recognition models will integrate algorithms designed to effectively filter out background noise. Furthermore, developers are increasingly focusing on training models with diverse datasets that include various accents and dialects. This can lead to a more inclusive and accurate understanding of spoken language.

Avatar photo
Written By

Dr. Tim Sandle is Digital Journal's Editor-at-Large for science news. Tim specializes in science, technology, environmental, business, and health journalism. He is additionally a practising microbiologist; and an author. He is also interested in history, politics and current affairs.

You may also like:

World

US Secretary of State Antony Blinken (L) is paying his second visit to China in less than a year - Copyright POOL/AFP Mark SchiefelbeinShaun...

Business

Google-parent Alphabet soared with Microsoft in after-hours trade following forecast-beating earnings - Copyright GETTY IMAGES NORTH AMERICA/AFP Drew AngererMarkets were mixed on Friday after...

Life

An expert explains why keen gamers should consider running as part of their regular routine.

World

People wave the Palestinian flag during protests in Doha after the outbreak of the Gaza war - Copyright AFP Rabih DAHERCallum PATONCriticism of Qatar...