Rosenverse

Accessible only to conference ticket holders.

Log in Create account Buy conference recordings

For 90 days after a conference, only paid ticket holders can watch conference videos. After that, all Gold members have access.

If you have purchased recording access and cannot see the video, please contact support.

Human vs. machine: Testing AI’s ability to synthesize and analyze research
Conference ticket
Wednesday, March 11, 2026 • Advancing Research 2026
Share the love for this talk
Human vs. machine: Testing AI’s ability to synthesize and analyze research
Speakers: Laura Klein
Link:

Summary

Nielsen Norman Group (NNG) has conducted and continues to conduct extensive research testing various large language model (LLM) tools designed for research synthesis and analysis. Our goal was to determine whether these AI-powered tools could meaningfully accelerate the work of experienced UX researchers. Through rigorous testing across multiple models and specialized research tools, we’ve found that while a few tools provide modest speed improvements for experienced researchers, none come close to replacing human expertise in research synthesis and analysis. The core problem is that these tools consistently exhibit critical flaws: they hallucinate findings, fail to identify meaningful patterns in qualitative data, cannot adequately consider nuanced research questions, and produce only superficial, high-level summaries of participant behavior. What makes this particularly dangerous is that these AI-generated outputs often have the veneer of legitimate research results—they look professional and sound plausible. However, closer inspection reveals significant gaps, inaccuracies, and missed insights that would mislead stakeholders and result in poor design decisions. The appearance of competence masks fundamental limitations that make these tools unreliable for serious research work. While we’ve found several places in the research process that can benefit from LLM usage, analysis and synthesis consistently falls short. In this talk, I can share the specific research we’re doing and explain what actually works.

Key Insights

  • AI tools frequently produce insight-shaped outputs but often lack the rigor and accuracy of trained human researchers.

  • AI moderators cannot currently assess user behavior beyond spoken words, missing key usability observations like failed or inefficient tasks.

  • Contextual elements such as environmental interruptions are critical in research but are invisible to AI tools.

  • Synthetic users generated by AI tend to produce overly positive, unrealistic feedback that can mislead product teams.

  • AI excels at finding semantic connections and grouping codes in large, already coded qualitative datasets quickly.

  • Meta-analysis of large repositories using AI can uncover recurring user themes, like change aversion, much faster than manual methods.

  • Integrating AI with organizational systems to pull in diverse data sources improves context but requires expert setup and is not yet simple.

  • AI’s context window limitations cause it to forget earlier input, affecting the accuracy of multi-turn interactions.

  • Even trained researchers must use AI outputs cautiously, vetting insights to maintain research quality.

  • Effective user research depends on human synthesis, collaboration, and contextual understanding, areas where AI currently fails.

Notable Quotes

"AI can generate insights, but it does not do them as well as a moderately trained human researcher."

"There is a world of difference between what a participant says and what they actually do, and AI misses that completely."

"AI tells you what you want to hear, which is dangerous if you’re making product decisions based on synthetic feedback."

"Our job as researchers is not making reports or interviewing users; it’s providing actionable, correct insights."

"AI tools are incentivized to produce final deliverables, but that’s an output, not the essence of research."

"AI is pretty good at finding semantic patterns among codes after human researchers have done the initial coding."

"Nobody is going to be satisfied by insight-shaped answers or high-level summaries masquerading as breakthroughs."

"AI cannot notice body language, tone, or environmental context during a research session."

"Using AI to scan large archives of research is a game changer for meta-analyses, even if it’s imperfect."

"Well-set-up AI systems pulling data from multiple company sources will have more context, but it’s still limited compared to human understanding."

Ask the Rosenbot
Melissa Schmidt
How UX Research Hit It Big in Las Vegas
2019 • Enterprise Experience 2019
Gold
Dharani Perera
The mandala of service design: unlocking alignment and action through service design
2025 • Advancing Service Design 2025
Gold
Laura Gatewood
Beyond Buzzwords: Adding Heart to Effective Slack Communication
2024 • DesignOps Summit 2024
Gold
Prayag Narula
Empowering Designers to do Good Research
2022 • Advancing Research 2022
Gold
Uday Gajendar
Leading through the long tail of trauma
2022 • Advancing Research Community
Billy Carlson
Principles of Team Wireframing
2023 • DesignOps Summit 2023
Gold
Chris Hammond
Embedding sustainability into enterprise design and development: A journey towards "sustainability consciousness"
2025 • Climate UX Interest Group
Sam Proulx
Accessibility: An Opportunity to Innovate
2022 • Advancing Research 2022
Gold
Nalini P. Kotamraju
Two Jobs in One: Being a “Leader who is a Researcher” and a “Researcher who is a Leader"
2021 • Advancing Research 2021
Gold
Alexis Lucio
Scaling Accessibility Through Design Systems
2022 • Design at Scale 2022
Gold
Kevin Bethune
Gatekeepers and Servant Leadership
2020 • DesignOps Community
Christian Crumlish
The Pygmalion Effect: In Which a Vibe Coding Experiment Becomes a Million Lines…
2025 • Rosenfeld Community
Jilanna Wilson
Distributed DesignOps Management
2019 • DesignOps Community
Jake Burghardt
Stop wasting research: Unlock more value from research insights
2025 • Rosenfeld Community
Bria Alexander
Opening Remarks
2021 • Design at Scale 2021
Gold
Sam Ladner
Data Exhaust and Personal Data: Learning from Consumer Products to Enhance Enterprise UX
2016 • Enterprise UX 2016
Gold

More Videos

Scott Jensen

"It’s not move fast and break things; it’s slow and steady."

Scott Jensen Sarah Delaney Carmen Liu

Short Take #2: UX/Product Lessons from Your Industry Peers

December 6, 2022

Theresa Slate

"Give three pieces of evidence for every piece of feedback you provide an individual."

Theresa Slate Erin Robertson

Why Changing Hearts & Minds Doesn’t Work When Promoting DE&I Efforts, but Checklists Do

October 4, 2023

Steve Portigal

"The in-house researcher has some elements of the client and stakeholder roles."

Steve Portigal

Looking Back…to Look Ahead

March 26, 2024

Kristin Skinner

"Personalization is not just a trend; it’s a necessity."

Kristin Skinner

Five Years of DesignOps

September 29, 2021

Ned Dwyer

"It’s not always a charismatic repository that people flock to, but tools that seek you out where you work are more effective."

Ned Dwyer Emily Stewart James Wallis

The Intersection of Design and ResearchOps

September 24, 2024

Nicole Bergstrom

"If you want to work with Hilton, you have to bring accessibility to the table like security and privacy."

Nicole Bergstrom Anna Cook Kate Kalcevich Saara Kamppari-Miller

AccessibilityOps: Moving beyond “nice to have”

September 19, 2024

Christian Rohrer

"The golden trapezoid of user research combines quantitative behavioral and attitudinal data with field studies for best insights."

Christian Rohrer

Insight Types That Influence Enterprise Decision Makers

May 13, 2015

Bria Alexander

"If you want to avoid time zone confusion, you can switch the conference schedule to your local time on the program page."

Bria Alexander

Opening Remarks

November 17, 2022

Cassini Nazir

"We’re not feeling what others feel, we’re feeling what we think they’re feeling in our own way."

Cassini Nazir Meah Lin

The Dangers of Empathy: Toward More Responsible Design Research

March 27, 2023