Connect with us

Hi, what are you looking for?

Tech & Science

Are we falling out of love with voice assistants?

Users express dismay when interacting with smartphones, where frustration often arises due to misunderstandings between humans and their devices.

Amazon Echo speaker configured for the Alexa voice assistant. Image © Tim Sandle.
Amazon Echo speaker configured for the Alexa voice assistant. Image © Tim Sandle.

Voice assistants have stalled a little in terms of their development arc, with consumers commonly expecting more sophisticated algorithms. For example, 41 percent of voice assistant users have reported experiencing miscommunication issues with their devices every week.

This represents part of a data set compiled by the technology firm Cubicfox.

In a world becoming more reliant on voice recognition, clarity is crucial. Thomas M. Pentz, CEO of Cubicfox explains to Digital Journal: “The future of communication lies not just in the technology, but in our ability to understand and adapt to it.”

In particular, users express dismay when interacting with smartphones, where frustration often arises due to misunderstandings between humans and their devices.

Understanding the Gap

According to a report from Microsoft, smartphone users are increasingly frustrated with their devices’ ability to comprehend their voice commands. 41 percent of users express worries regarding trust, privacy, and passive listening. This underscores a discrepancy between user expectations and the present capabilities of speech recognition technology.

Reasons for misunderstandings

The reason  why a voice assistant sometimes fails to deliver an intelligible response include:

Background noise: Software algorithms responsible for voice recognition rely on clear audio input. When background noise creeps in, these algorithms face difficulty separating your voice from the surrounding sounds.

Accents and dialects: According to Frontier voice assistants like Alexa achieve a 55 percent accuracy rate for native speakers. However, this drops significantly for non-native speakers or those with strong accents.

Ambiguous phrasing: The way we naturally speak often involves incomplete sentences, slang, or informal language. These elements can be challenging for AI models to interpret compared to grammatically correct and formal phrasing.

A path forwards is with improved noise cancellation. It is likely that newer speech recognition models will integrate algorithms designed to effectively filter out background noise. Furthermore, developers are increasingly focusing on training models with diverse datasets that include various accents and dialects. This can lead to a more inclusive and accurate understanding of spoken language.

Avatar photo
Written By

Dr. Tim Sandle is Digital Journal's Editor-at-Large for science news. Tim specializes in science, technology, environmental, business, and health journalism. He is additionally a practising microbiologist; and an author. He is also interested in history, politics and current affairs.

You may also like:

Tech & Science

Digital Journal announced as official media partner for Innovation Week in Calgary.

Life

Internet personality Guilherme Werner chatted about representing Brazil in the 2024 Mister Universe competition.

Tech & Science

Awareness and proactive measures are the best defences against the evolving tactics of cyber attackers.

Business

The skills gap in Site Reliability Engineering (SRE) talent is significantly impacting businesses that are adopting cloud-native architectures.