Log in or create a free Rosenverse account to watch this video.
Log in Create free account100s of community videos are available to free members. Conference talks are generally available to Gold members.
Summary
Let’s face it, many of us feel daunted by statistics. But we also know that colleagues and clients ask whether our research has “statistically significant” results. Erin’s book Design for Impact helps you to test your hypotheses about improving design, and she guides you through deciding on your effect sizes to help you get those statistically significant results. Caroline’s book Surveys That Work talks about “significance in practice” and she’s not all that convinced about whether it’s worth aiming for statistical significance. Watch this lively session where Erin and Caroline compared and contrasted their ideas and approaches - helped by your questions and contributions.
Key Insights
-
•
Statistical significance often confuses practitioners because it requires mentally flipping hypotheses and disproving nulls, which is cognitively demanding.
-
•
Effect size is critical to understanding whether a change detected by statistics is meaningful in practice, a concept often neglected in statistics education.
-
•
Fast progress isn't necessarily good progress; teams benefit from slowing down and using statistics to ensure they're moving in the right direction.
-
•
Engineers can be reluctant to implement experiments due to the extra coding load, but they respond well when they understand the learning value gained.
-
•
Survey results (the numerical outcomes) are often confused with the number of respondents required for statistical significance, leading to misunderstandings.
-
•
Statistical thresholds like 95% confidence can be adjusted depending on project needs; lower confidence levels are sometimes acceptable.
-
•
A good hypothesis often starts as an intuitive guess, which gets refined over time through repeated testing and data collection.
-
•
Qualitative and quantitative research should be viewed as complementary tools in a holistic research approach rather than opposed methods.
-
•
AI tools can help generate first drafts of survey questions, but human-centered pilot testing is essential to avoid errors and misinterpretations.
-
•
It's common and acceptable to act on results that are significant in practice but not statistically significant, especially when outcomes clearly affect users.
Notable Quotes
"Statistics is hard because you have to flip flop in your head: think of a hypothesis, then a null hypothesis, then try to disprove the null."
"People confuse statistical significance with significance in practice — they want to know if the change is meaningful, not just mathematically significant."
"Fast is not a virtue in and of itself; moving slower and acting with intention ensures you go in the right direction."
"Engineers hate writing more code, so getting them to buy into experiments means showing the value of the learning on the other side."
"You can have an effect size that matters in practice but isn’t statistically significant, like five users failing a key task in usability testing."
"Most science starts with somebody pulling a number out of their ass — it’s okay to start with a gut instinct or guess."
"Statistics is another tool in our toolbox, part of a hierarchy of evidence that includes qualitative and quantitative methods."
"AI can create first drafts of survey questions, but unless you pilot test with real humans, you won’t know if your audience gets it."
"A lot of people think 95% confidence is the only way, but you can adjust confidence levels based on your situation and needs."
"Start with basics like means, minimums, and ranges — statistics rapidly becomes less mysterious and more useful with practice."
Or choose a question:
More Videos
"You want to flex one or two variables to create range scenarios—best case, middle case, and worst case."
Giff ConstableFinancial fluency for product leaders: AMA with Giff Constable
April 11, 2024
"Users frequently felt overwhelmed and misunderstood; this was a significant barrier to a positive shopping experience."
Kritika YadavOptimizing AI Conversations: A Case Study on Personalized Shopping Assistance Frameworks
June 10, 2025
"Deep listening means hearing not just what is said but what isn’t said, especially for people with different language or cultural backgrounds."
Alla WeinbergHow to Build and Scale Team Safety
January 8, 2024
"Persuasive technology shifted design focus from ease of use to influencing what users do."
Harry Brignull Mark Leiser Robert StribleyBeyond Clicks and Tricks: Why deceptive design has grown into a regulatory faultline
January 16, 2026
"Companies with high corporate design maturity perform better financially and have happier employees."
Kit UngerTheme 2: Introduction
June 10, 2021
"Introducing change is introducing stress and fear and feelings of rejection."
Sofia QuinteroBeyond Tools: The Messy Business of Implementing Research Repositories
March 10, 2022
"When you have a dedicated designer, you see faster learning, iteration, collaboration, and agility."
Jeff GothelfThe Intersection of Lean and Design
January 10, 2019
"She earned the title best boss ever by being a darn good human being."
Kit Unger Lada GorlenkoTheme 3 Intro
June 10, 2022
"I resist premature optimization of tagging and structure; I want to be 80% there rather than perfect and distracted."
Jorge ArangoThe Best of Both Worlds: How to Integrate Paper and Digital Notes (1st of 3 seminars)
April 5, 2024
Latest Books All books
Dig deeper with the Rosenbot
What are the challenges and opportunities posed by democratization of user research within organizations?
How many books are available in the Rosenverse Gold subscription, and in what format?
What are the benefits of linking AI-powered research repositories with tools like Slack for organizational adoption?