Log in or create a free Rosenverse account to watch this video.
Log in Create free account100s of community videos are available to free members. Conference talks are generally available to Gold members.
Summary
Imagine slipping on a sleek pair of smart glasses. Not only do you look sharp, the glasses capture everything you see, hear, and do. Your AI assistant—built into the glasses and synced to your email, social media accounts, health apps, and finances—manages your life. It’s tasked with paying bills, booking trips, replying to messages, even helping you swipe right. Over time, you find yourself chitchatting with your AI assistant. You call him Charlie. Now imagine you’re a threat actor. That trust between user and AI assistant? It’s your entry point. If your product is powered by AI, you’re not just designing features—you’re designing an entire relationship. You’re designing Charlie. Let’s talk about where that goes wrong—and how to get it right.
Key Insights
-
•
Users often do not understand why AI-powered systems request extensive personal data, increasing privacy risks.
-
•
Trust in AI agents can become excessive, creating new vectors for manipulation by threat actors.
-
•
Security issues typically occur beneath the surface until alerts disrupt the user experience, often causing frustration.
-
•
Prompt injection attacks pose a novel threat where malicious inputs manipulate AI agents to access sensitive user data.
-
•
Multimodal AI interfaces introduce complexity in security decisions, increasing chances for user errors.
-
•
Secure by default settings reduce burden on users and improve overall protection without requiring user intervention.
-
•
Cross-disciplinary collaboration between UX, security, product, legal, and compliance teams is crucial for safer AI design.
-
•
Users need clear, contextual guidance during onboarding to make informed decisions about data sharing and security settings.
-
•
Transparency about AI limitations and giving users the option to reverse AI actions are essential for building trust.
-
•
Threat actors are likely to exploit growing AI access to personal data and automate vulnerabilities discovery.
Notable Quotes
"When a product is powered by AI, you're not just designing the features; you are designing an entire relationship."
"Charlie is like the most annoying coworker who constantly surfaces problems but never offers solutions to Alice."
"Threat actors probably know your system better than you do and are looking for any entry points to exploit."
"Alice often perceives Charlie as just another barrage of alerts filled with jargon she doesn't understand."
"Prompt injection attacks can trick AI agents into accessing private data like emails without the user realizing."
"People become incrementally more comfortable giving away data because they see the value AI provides."
"We need secure defaults that protect users out of the box without them having to figure it out."
"Alert fatigue is real; users can't be burdened with constant security decisions or they'll ignore them."
"Giving users the ability to reverse AI-driven actions is critical but currently underexplored."
"If Charlie has been tampered with, Alice needs a clear way to be alerted that she shouldn't trust it."
Or choose a question:
More Videos
"We started to reference something that became to call the triple D challenges: duplicated efforts, disjointed workflows, and distance from users."
Sean Fitzell Sarah Han Kayla FarrellCraft of User Research: Building Out Jobs to be Done Maps
March 12, 2021
"Most former extremists don’t come up with the idea of involvement themselves, somebody always approaches them."
Jim KalbachPeace is waged with sticky notes: Mapping Real-World Experiences
June 14, 2018
"Companies know they’re doing wrong but don’t change because incentives don’t align or they lack operational competence."
Cennydd BowlesResponsible Design in Reality
June 9, 2021
"Our KSPs do not change year over year because changing them frequently is disruptive for the business direction."
Saara Kamppari-Miller Nicole Bergstrom Shashi JainKey Metrics: Comparing Three Letter Acronym Metrics That Include the Word “Key”
November 13, 2024
"Manuel Herrera is creating sketch notes for every session; you’ll have access to those."
Uday Gajendar Louis RosenfeldDay 1 Welcome
June 4, 2024
"The title is becoming less important. Focus on the outcome you want to create and whether it feels meaningful."
Sara ConklinExit Interview: 20 Years of Tech, One Very Big Bet, and a Lot of Heat Pumps
April 10, 2026
"Taking stakeholders on the journey by involving them in interviews really changed how they saw the product and business strategy."
Dr Chloe SharpUsing Evidence and Collaboration for Setting and Defending Priorities
November 29, 2023
"To be business and design, not business versus design, we have to change the language we use to talk about value."
Nathan ShedroffRedefining Value: Bridging the Innovation Culture Divide
May 14, 2015
"This is probably the biggest change we are going to see within our working careers."
Robert Fabricant Sahibzada Mayed Nidhi Singh RathoreIndustry junctures: Paths forwards for UXR and the critical decisions that get us there [Advancing Research Community Workshop Series]
October 2, 2024
Latest Books All books
Dig deeper with the Rosenbot
How many books are available in the Rosenverse Gold subscription, and in what format?
How can organizations manage the noise created by non-expert researchers conducting user research?
What does a successful healthcare UX career look like in terms of accumulating influence and aligning with clinical/business goals?