The trust and privacy concerns no one is talking about
In an age where voice-activated technology is ubiquitous, the question on many consumers’ minds is: are virtual assistants like Alexa and Google Assistant really looking out for us—or are they spying on us?
According to our AI Impact Report, commissioned by House 337, a significant 39% of consumers express trust issues with Generative AI technologies, with a notable concern about privacy that’s equally prevalent across various platforms, including voice assistants.
The hesitation surrounding these tools can largely be attributed to fears regarding data collection, storage, and usage—over half of respondents (55%) reported worries in this area. For those aged 55 and over, these concerns were even more pronounced, with 48% expressing distrust specifically in the capabilities of Generative AI.
What can brands do to address these fears?
First, transparency is key. Companies must clearly communicate how they collect and use data to foster a sense of security among users. Implementing robust security measures and openly discussing these practices can significantly bolster consumer confidence. The rise of AI technologies presents tremendous opportunities, but the path forward hinges on building a foundation of trust.