Rosenverse

Accessible only to conference ticket holders.

Log in Create account Buy conference recordings

For 90 days after a conference, only paid ticket holders can watch conference videos. After that, all Gold members have access.

If you have purchased recording access and cannot see the video, please contact support.

Human vs. machine: Testing AI’s ability to synthesize and analyze research

Conference ticket
Wednesday, March 11, 2026 • Advancing Research 2026
Share the love for this talk
Human vs. machine: Testing AI’s ability to synthesize and analyze research
Speakers: Laura Klein
Link:

Summary

Nielsen Norman Group (NNG) has conducted and continues to conduct extensive research testing various large language model (LLM) tools designed for research synthesis and analysis. Our goal was to determine whether these AI-powered tools could meaningfully accelerate the work of experienced UX researchers. Through rigorous testing across multiple models and specialized research tools, we’ve found that while a few tools provide modest speed improvements for experienced researchers, none come close to replacing human expertise in research synthesis and analysis. The core problem is that these tools consistently exhibit critical flaws: they hallucinate findings, fail to identify meaningful patterns in qualitative data, cannot adequately consider nuanced research questions, and produce only superficial, high-level summaries of participant behavior. What makes this particularly dangerous is that these AI-generated outputs often have the veneer of legitimate research results—they look professional and sound plausible. However, closer inspection reveals significant gaps, inaccuracies, and missed insights that would mislead stakeholders and result in poor design decisions. The appearance of competence masks fundamental limitations that make these tools unreliable for serious research work. While we’ve found several places in the research process that can benefit from LLM usage, analysis and synthesis consistently falls short. In this talk, I can share the specific research we’re doing and explain what actually works.

Key Insights

  • AI tools frequently produce insight-shaped outputs but often lack the rigor and accuracy of trained human researchers.

  • AI moderators cannot currently assess user behavior beyond spoken words, missing key usability observations like failed or inefficient tasks.

  • Contextual elements such as environmental interruptions are critical in research but are invisible to AI tools.

  • Synthetic users generated by AI tend to produce overly positive, unrealistic feedback that can mislead product teams.

  • AI excels at finding semantic connections and grouping codes in large, already coded qualitative datasets quickly.

  • Meta-analysis of large repositories using AI can uncover recurring user themes, like change aversion, much faster than manual methods.

  • Integrating AI with organizational systems to pull in diverse data sources improves context but requires expert setup and is not yet simple.

  • AI’s context window limitations cause it to forget earlier input, affecting the accuracy of multi-turn interactions.

  • Even trained researchers must use AI outputs cautiously, vetting insights to maintain research quality.

  • Effective user research depends on human synthesis, collaboration, and contextual understanding, areas where AI currently fails.

Notable Quotes

"AI can generate insights, but it does not do them as well as a moderately trained human researcher."

"There is a world of difference between what a participant says and what they actually do, and AI misses that completely."

"AI tells you what you want to hear, which is dangerous if you’re making product decisions based on synthetic feedback."

"Our job as researchers is not making reports or interviewing users; it’s providing actionable, correct insights."

"AI tools are incentivized to produce final deliverables, but that’s an output, not the essence of research."

"AI is pretty good at finding semantic patterns among codes after human researchers have done the initial coding."

"Nobody is going to be satisfied by insight-shaped answers or high-level summaries masquerading as breakthroughs."

"AI cannot notice body language, tone, or environmental context during a research session."

"Using AI to scan large archives of research is a game changer for meta-analyses, even if it’s imperfect."

"Well-set-up AI systems pulling data from multiple company sources will have more context, but it’s still limited compared to human understanding."

Ask the Rosenbot
Jessamyn Edwards
Surviving Your UX Career in Enterprise Design
2021 • Enterprise Community
Sam Proulx
Online Shopping: Designing an Accessible Experience
2023 • Enterprise UX 2023
Gold
Harry Max
Failure Friday #5: Lessons from a SaaS Design Failure
2025 • Rosenfeld Community
Michael Polivka
Scaling Design through Relationship Maps
2017 • DesignOps Summit 2017
Gold
Wendy Johansson
Design at Scale: Behind the Scenes
2021 • Enterprise Community
Patrizia Bertini
The (r)evolution of designOps: It’s Time to Think (really) BIG
2025 • DesignOps Summit 2025
Gold
Bria Alexander
Opening Remarks
2023 • DesignOps Summit 2023
Gold
Sheryl Cababa
Expanding Your Design Lens with Systems Thinking
2023 • Enterprise Community
Jemma Ahmed
Redefining the research toolkit: Expanding methodologies for a changing world
2025 • Advancing Research 2025
Gold
Laura Klein
Unique challenges of innovation in enterprises
2020 • Enterprise Community
Discussion
2017 • Enterprise Experience 2017
Gold
Denise Jacobs
Interactive Keynote: Social Change by Design
2024 • Enterprise Experience 2020
Gold
Yalenka Mariën
Designing for Digital Inclusion in the Belgian Government
2021 • Civic Design 2021
Gold
John Cutler
Prioritization for designers and product managers (1st of 3 seminars)
2024 • Rosenfeld Community
Erin Hoffman-John
This Game is Never Done: Design Leadership Techniques from the Video Game World
2017 • DesignOps Summit 2017
Gold
Gordon Ross
12 Months of COVID-19 Design and Digital Response with the British Columbia Government
2021 • Civic Design 2021
Gold

More Videos

Jon Fukuda

"AI won’t transform all things at once; it’s about piloting and delivering discrete business value first."

Jon Fukuda Amy Evans Ignacio Martinez Joe Meersman

The Big Question about Innovation: A Panel Discussion

September 25, 2024

Sam Proulx

"Accessibility is not a shackle; it’s a way to expand our minds and innovate."

Sam Proulx

Accessibility: An Opportunity to Innovate

March 9, 2022

Anna Avrekh

"If you don’t invest in retention, you get the leaky bucket syndrome where diverse talent keeps leaving."

Anna Avrekh Amy Jiménez Márquez Morgan C. Ramsey Catarina Tsang

Diversity In and For Design: Building Conscious Diversity in Design and Research

June 9, 2021

Greg Nudelman

"Bots is like a perfect on-ramp for UX for AI."

Greg Nudelman

Designing Conversational Interfaces

November 14, 2019

George Abraham

"The biggest gain is about getting value out sooner, with an 80 percent efficiency improvement from design to running app."

George Abraham Stefan Ivanov

Design Systems To-Go: Reimagining Developer Handoff, and Introducing App Builder (Part 2)

October 1, 2021

Shipra Kayan

"Research insights should belong to the team, not just the researcher."

Shipra Kayan

Make your research synthesis speedy and more collaborative using a canvas

January 24, 2025

Sam Proulx

"It is very difficult for somebody who has never used assistive technology to have the same experience as someone who uses it daily for everything."

Sam Proulx

Designing For Screen Readers: Understanding the Mental Models and Techniques of Real Users

December 10, 2021

Shipra Kayan

"Our one metric that mattered was maximizing the number of VOС-tagged tickets solved – showing customer feedback made it to the roadmap."

Shipra Kayan

How we Built a VoC (Voice of the Customer) Practice at Upwork from the Ground Up

September 30, 2021

Dane DeSutter

"Look at your old data again with a new lens; reanalyzing can save money and be less extractive."

Dane DeSutter

Keeping the Body in Mind: What Gestures and Embodied Actions Tell You That Users May Not

March 26, 2024