Log in or create a free Rosenverse account to watch this video.
Log in Create free account100s of community videos are available to free members. Conference talks are generally available to Gold members.
Summary
Deceptive design has a long prehistory. Before it became a recognised field with legal and regulatory consequences, it lived quietly in user interfaces, persuasive patterns, and the small frictions that nudged people into choices they never quite meant to make. This fireside chat brings together three leading voices to trace that evolution and map its next steps. Dr Harry Brignull opens with the historical and cultural arc of deceptive design, showing how early interface tricks solidified into a taxonomy and a movement. Author Robert Stribley then guides the conversation into the world of privacy-protective design, examining why users often remain indifferent to risks, why organisations struggle to prioritise privacy, and how better design principles can restore agency rather than erode it. Dr Mark Leiser closes by shifting from screens to systems. His work reveals how dark patterns now extend far beyond the UI, emerging in algorithmic optimisation, platform architecture, and AI-driven inference. This is where deceptive design becomes a structural problem, not a cosmetic one, and where law, regulation, and system design collide. Together, Brignull, Stribley, and Leiser explore how design, privacy, and digital regulation are becoming inseparable, and what it will take to build technologies that respect people rather than manipulate them. The result is an interdisciplinary, future-facing conversation about one of the most urgent challenges in the digital environment today.
Key Insights
-
•
Dark patterns originated from early attempts to influence behavior beyond usability, gaining wide attention after Harry's 2010 talk.
-
•
Mark's legal perspective frames deceptive design as causing tangible financial harm, distinct from traditional privacy issues.
-
•
The GDPR and similar regulations struggle with enforcement due to lack of precise guidance and intersecting consumer and privacy laws.
-
•
Designers often lack legal training, leading to reliance on compliance approvals rather than proactive ethical or legal considerations.
-
•
Deceptive UX patterns like manipulative cookie banners offer false choices that obscure users’ ability to reject tracking.
-
•
Systemic deceptive architectures use AI to adapt prompts based on user behavior, presenting a new, harder-to-detect form of manipulation.
-
•
There is a pressing need for design industry groups to set professional standards against manipulative or coercive design practices.
-
•
Consent models like presumed consent in the U.S. mask the reality that users rarely provide informed consent.
-
•
The privacy paradox—users claiming to value privacy but behaving otherwise—may be explained by deceptive design trickery.
-
•
Recourse systems in digital products are underdeveloped, leaving users powerless when harmed beyond the interface.
Notable Quotes
"Everyone knew tech was changing the world and we all had this peak at how amazing the web was going to be."
"User research then focused on basic usability problems like removing reset buttons that wiped forms."
"Persuasive technology shifted design focus from ease of use to influencing what users do."
"I got invited to be an expert witness on an FTC legal case about deceptive designs, which changed my career."
"From a legal point of view, taking money from people without their true consent is a tangible harm."
"Designers often say 'has it been approved by legal and compliance? Great, let's not worry about privacy anymore.'"
"Deceptive patterns in cookie banners give users false choices that make it harder to reject tracking."
"Systemic deceptive design is harder to regulate because manipulation happens across time and user context."
"Designing for privacy is like designing for accessibility — you should aim beyond mere compliance to user-centered solutions."
"We have too many privacy lawyers and not enough consumer lawyers, which clouds enforcement priorities."
Or choose a question:
More Videos
"The peer review process is slow and gatekeeps change, so innovating from the margins is strategic."
Kristen Honey"Let’s Talk About Data and Crisis”: Public Digital Service Delivery = Open Data + Human Centered Design
November 18, 2021
"Hallucinations are a kind of AI creativity—dreaming, invention, fiction—all forms of hallucination."
Matt WebbContext Window: Five Futures for AI
June 11, 2025
"You come from the future. Now go build it."
Dana ChisnellThe Sensemaking Business
March 10, 2026
"Researchers are shy about sharing work because they worry about being judged — peer and management support help."
Kate Towsey Jake BurghardtResearchOps AMA with Kate Towsey & Jake Burghardt
October 16, 2025
"I’m trying to separate who I am from what I produce and the roles I hold."
Ashley SewallExit Interview #5: Designing My Life After Tech
February 19, 2026
"Listen hard, change fast."
Jacqui Frey Dan WillisPanel Discussion: Integrating DesignOps
November 7, 2018
"I may be ruining his experience because I’m forgetting it’s his experience, not mine."
Iram ShahClosing Keynote: The View from the Top
June 4, 2019
"The LAUSD stakeholder felt armed with data and designs to articulate future benefits in a language the school board could understand and quantify."
JD BuckleyCommunicating the ROI of UX within a large enterprise and out on the streets
June 14, 2018
"AI isn’t about being an expert for most roles yet, but being AI conversant and experimental is crucial."
Dave Hoffer Joanne WeaverUX Job Search AMA #3 with Joanne Weaver and Dave Hoffer
July 16, 2025