Log in or create a free Rosenverse account to watch this video.
Log in Create free account100s of community videos are available to free members. Conference talks are generally available to Gold members.
Summary
Enthusiasm for AI tools, especially large language models like ChatGPT, is everywhere, but what does it actually look like to deliver large-scale user-facing experiences using these tools in a production environment? Clearly they're powerful, but what do they need to make them work reliably and at scale? In this session, Sarah provides a perspective on some of the information architecture and user experience infrastructure organizations need to effectively leverage AI. She also shares three AI experiences currently live on Microsoft Learn: An interactive assistant that helps users post high-quality questions to a community forum A tool that dynamically creates learning plans based on goals the user shares A training assistant that clarifies, defines, and guides learners while they study Through lessons learned from shipping these experiences over the last two years, UXers, IAs, and PMs will come away with a better sense of what they might need to make these hyped-up technologies work in real life.
Key Insights
-
•
Most AI applications no longer require building foundation models from scratch; the focus is now on application development and integration.
-
•
Single, all-purpose chatbots (everything chatbots) are insufficient because they handle high ambiguity and diverse, often complex tasks poorly.
-
•
Sarah introduces the ambiguity footprint as a framework to measure AI application complexity and risks across several axes such as task complexity, context, interface, prompt openness, and sensitivity.
-
•
AI features that support simple, complimentary user tasks, rather than critical or complex ones, are easier and safer to build and scale.
-
•
Visible AI interfaces, like chatbots, set clearer user expectations but introduce more ambiguity and management overhead compared to invisible AI (e.g., keyboard optimizations).
-
•
Prompt engineering plays a crucial role in defining the boundaries of AI output, from very open-ended to highly restricted scopes.
-
•
Retrieval Augmented Generation (RAG) helps manage up-to-date context by dynamically querying relevant data chunks rather than using static corpus.
-
•
Evaluating AI outputs rigorously is essential but often underprioritized; without clear quality metrics, teams end up relying on subjective or anecdotal assessments.
-
•
Data ethics and distributed AI implementations can create blind spots, limiting feedback loops necessary for continuous AI model improvement.
-
•
Incrementally building AI applications with smaller ambiguity footprints helps organizations develop expertise and controls before tackling more complex, open-ended AI products.
Notable Quotes
"You’re not doing IA, but you’re always doing it."
"An everything chat bot is almost certainly not how you’re going to build it; realistically you’re building three apps in a trench coat."
"AI is ambiguous at best because we’re fully in the realm of probabilistic rather than deterministic programming."
"The more complex the task, the less likely it is to be successful with current AI."
"A task where AI adds a little something is honestly easier to get right than one where it’s absolutely critical."
"Visible AI interfaces introduce another place where you can add ambiguity."
"Retrieval Augmented Generation lets you supply specific relevant information to the model dynamically rather than everything at once."
"Evaluation might be the most important part of your entire development effort and is often the hardest to do well."
"You can’t just eyeball results and call it good; AI applications are expensive and complex and require systematic evaluation."
"Never build or buy an everything chat bot again; start with less ambiguous, targeted AI experiences."
Or choose a question:
More Videos
"We need to break the dream and move from spectators to actors who create knowledge grounded in our reality."
Verónica Urzúa Jorge MontielThe B-side of the Research Impact
March 12, 2021
"I can’t think of a single screen reader user who uses the default out-of-the-box configuration."
Sam ProulxUnderstanding Screen Readers on Mobile: How And Why to Learn from Native Users
March 27, 2023
"We often don’t prototype everything ourselves; instead, we role play with client team members to simulate complex organizational aspects."
Milan Guenther Benjamin KumpfThe $212 billion ‘so what?’: unlocking impact in development cooperation
November 20, 2025
"If you have a question for a speaker, put it in the thread for their talk, not in the general chat."
Bria AlexanderOpening Remarks
November 18, 2022
"Apple unlocked so much innovation by including accessibility in the phone with VoiceOver built-in."
Sam ProulxTo Boldly Go: The New Frontiers of Accessibility
June 10, 2022
"Dark mode is particularly useful in network and security operations centers where visibility in dark rooms is critical."
Theresa NeilJust Build Me a Dashboard!
April 9, 2019
"I've thrived on being wrong and making mistakes, even when I wasn't actually wrong."
Tricia WangSCALE: Discussion
June 15, 2018
"This was the first time the director saw the team doing something together."
Saara Kamppari-Miller"Prototype" vs "Prototype"--Breaking Down and Rebuilding Our Understanding of What We Do
October 24, 2019
"We have four themes over two days: systems thinking, information architecture and knowledge management, AI and machine learning, and designing the data."
Louis RosenfeldWelcome / Housekeeping
June 7, 2023
Latest Books All books
Dig deeper with the Rosenbot
What practical steps did Flight Center take to scale research participation without sacrificing insight quality?
How can teams integrate inclusive design into early product development stages to avoid costly late fixes?
Why is it important to consider disability as a dynamic state in inclusive design?