New

Now in Claude, ChatGPT, Cursor & more with our MCP server

Back to blog
Research12 min read

ChatGPT for User Research in 2026: What It Does Well, Where It Breaks, and the Purpose-Built Alternative

Researchers are using ChatGPT to write interview questions, analyse transcripts, and synthesise themes. Here is what works, what fails, and why a purpose-built AI research platform like Koji is the production-grade answer.

Koji Team

May 16, 2026

ChatGPT for User Research in 2026: What It Does Well, Where It Breaks, and the Purpose-Built Alternative

TL;DR: ChatGPT is a useful brainstorming partner for user research. It is not a safe production tool for analysing real customer transcripts, recruiting and moderating participants, or generating final stakeholder reports. The two biggest problems: (1) ChatGPT hallucinates verbatim quotes that do not exist in your transcripts, and (2) pasting customer data into a consumer LLM creates GDPR, HIPAA, and contractual confidentiality problems. Purpose-built AI research platforms like Koji handle the full workflow — AI-moderated interviews, structured questions, automatic thematic analysis, and one-click reports — under a research-grade data agreement.

In 2025, almost every researcher we talk to has tried using ChatGPT for at least one part of their user research workflow. The temptation is obvious: a single chat can draft discussion guides, cluster open-ended responses, and even write a stakeholder summary in seconds. The problem is that "seconds" hides multiple compounding failures that only surface when you check the work line-by-line against the original data.

This post is honest about what ChatGPT genuinely does well, where it breaks in ways researchers cannot afford, and how a purpose-built AI research platform avoids those failure modes.

Where ChatGPT genuinely helps

Let us start with what works. ChatGPT is a capable assistant for the pre-interview and brainstorming parts of a research project. Specifically:

  • Drafting discussion guides. Paste your research question and a target persona, ask for a 12-question semi-structured guide, and ChatGPT produces a solid first draft in under a minute. See how to write user interview questions for what to keep and what to cut from the output.
  • Translating questions. Need to run a study in three languages? ChatGPT handles initial translation; a native speaker still has to review for cultural framing.
  • Generating screener criteria. Describe your ideal participant; ChatGPT drafts screener questions and disqualification logic.
  • Brainstorming probe questions. Stuck on what to ask next? ChatGPT suggests 5–10 follow-up probes for any given user statement.
  • Summarising a single transcript you have already read. If you have read the transcript yourself, ChatGPT can produce a useful bullet summary you can verify against your own memory.

Used this way — as a thinking partner that a human always reviews — ChatGPT is a meaningful productivity gain. Researchers who treat it as scaffolding produce better discussion guides faster.

Where ChatGPT breaks for production research

The trouble starts when ChatGPT moves from "thinking partner" to "data analyst." Here are the failure modes documented in 2025–2026 research.

1. Fabricated quotes

The single most-cited problem in the qualitative-analysis literature: when researchers ask ChatGPT to identify themes and support them with verbatim quotes, the quotes are often synthesised rather than copied. The model produces a plausible-sounding sentence in the participant's voice that does not exist in the original transcript. This is the well-known LLM hallucination failure mode, and it is particularly dangerous in research because the entire credibility of qualitative work rests on quoting participants accurately.

A 2025 study by Uintent documented that LLM thematic analyses often invented or recombined participant statements, requiring researchers to verify every quote against line and segment numbers in the source transcript. That verification work erases the supposed time savings — and worse, researchers who don't verify ship reports with quotes the participant never said.

2. Compounding errors across long conversations

Research from the International Journal of Qualitative Methods (Morgan, 2023) and later studies confirm that ChatGPT hallucinates significantly more often in later rounds of a long chat because the model conditions on its own previous errors. One mistake early in a thematic-analysis session can propagate through every subsequent prompt, producing a final report with stacked inaccuracies. A 2026 OpenAI study found that 3–20% of incorrect references reappear in later turns of the same conversation.

This is fatal for research, where the analyst typically asks dozens of clarifying questions across a single chat session.

3. GDPR, HIPAA, and confidentiality

If you paste real customer transcripts into the consumer ChatGPT (Free or Plus), you have transferred personally identifiable information (PII) to a US-based processor. Under GDPR this is a controlled cross-border transfer that almost certainly requires explicit consent your participants did not give. Under HIPAA, only ChatGPT Enterprise with a signed Business Associate Agreement (BAA) is compliant — Free and Plus are not. For B2B research with enterprise customers, your contractual confidentiality terms almost certainly forbid this kind of transfer. We covered the legal nuance in can I paste user interviews into ChatGPT — a guide to GDPR and LLMs.

Researchers who paste interview data into the consumer ChatGPT are creating real, demonstrable risk for their company.

4. No moderation capability

ChatGPT cannot conduct an interview with a participant. It can suggest questions, but it cannot run a 15-minute conversation, probe vague answers in real time, follow a discussion-guide branching logic, record audio, or produce a transcript. Every interview in a ChatGPT-only workflow still requires a human moderator on a video call. That is the most expensive part of qualitative research — and ChatGPT does not touch it.

5. No recruitment

ChatGPT does not have a participant panel, scheduling, screening logic, or incentive payouts. You will still need participant recruitment platforms, a calendaring tool, and a screener.

6. No structured reporting workflow

ChatGPT will produce a report when you ask, but it does not version control it, does not let your stakeholders comment on it, and does not stay in sync as you add new interviews. Every report is a one-off Markdown blob you have to paste into Notion or Google Docs by hand.

The compounding cost of "free"

The argument for ChatGPT is "it's free or $20/month." Stack up the hidden costs:

  • Verification time — checking every quote against line numbers in the source transcript can take 2–4 hours per study.
  • Legal review — if your data crosses GDPR or HIPAA boundaries, your legal team is involved.
  • Risk of incorrect reports — a single fabricated quote in a stakeholder deck destroys research credibility for the year.
  • Moderator time — every interview still requires a human moderator.
  • Recruiting cost — you still pay panel providers.
  • No async scale — you cannot run 50 interviews in parallel; you can only chat with the model.

For a single small study, ChatGPT looks free. For a quarterly research program, it is substantially more expensive than a purpose-built platform — once you count the human hours and risk.

What a purpose-built AI research platform does differently

Koji was built specifically for the failure modes above. Here is how the architecture differs from "I paste transcripts into ChatGPT":

Quote fidelity

Koji's thematic analysis is constrained to the actual transcript — every quote in the final report links back to the segment in the source recording where it appears. No fabrication is possible because the system retrieves verbatim quotes rather than generating them.

Voice moderation, not just analysis

Koji runs the interview itself with an ElevenLabs voice AI moderator that asks your discussion-guide questions, probes vague answers in real time, and adapts to the participant's pace. You do not have to be on the call. Participants take the interview on their own schedule from a shared link.

Structured + open in the same session

Koji blends 6 structured question types — open-ended, scale, single-choice, multiple-choice, ranking, yes/no — with AI follow-ups in a single session. You get clean quantitative comparability plus qualitative depth from one study. ChatGPT cannot ask anyone anything.

Research-grade data handling

Koji is built for research workflows from day one. Participant consent is captured, recordings are stored with audit trails, and the data plane is designed for the contractual requirements of customer research — not the looser terms of a consumer chatbot.

Async scale

Ship a Koji study link to 100 participants and 100 interviews can complete in parallel over 48 hours. ChatGPT scales to one conversation per browser tab.

Reports your team can act on

One-click thematic reports cluster recurring patterns across all completed interviews with verbatim quotes, sentiment, and study-level summary. Versioned, shareable, and updated as more interviews complete.

A pragmatic hybrid workflow

The honest position: ChatGPT is fine for the early creative work; Koji is the production layer. A workflow that uses both well:

  1. Brainstorm in ChatGPT. Draft a discussion guide. Generate screener criteria. Brainstorm probe questions. Do not paste customer data.
  2. Set up the study in Koji. Configure the AI moderator with your discussion guide, add structured questions, and define your participant criteria.
  3. Run interviews in Koji. Share the interview link, import a participant CSV, or use participant recruitment platforms. Async, parallel, at scale.
  4. Analyse in Koji. Thematic clustering with verbatim-anchored quotes. No hallucination risk.
  5. Synthesise in your tool of choice. Export the report; bring it into Notion, Google Docs, or Linear for stakeholder distribution.

That is how a 2026 research function actually ships studies — not by pasting transcripts into a chatbot and hoping for the best.

The bigger picture: AI in research is not optional

Research from 2025 shows AI-powered research adoption increased 32% across UX and product teams, and ResearchOps automation is the fastest-growing line item in research budgets. Teams that pair purpose-built AI tools with traditional research craft are the ones scaling insight production in 2026.

The question is not "should I use AI for research?" — it is "which AI should I trust with my customer data, and where does it actually belong in my workflow?" Use ChatGPT for the parts of the job where hallucination and confidentiality do not matter. Use a purpose-built platform like Koji for the parts where they do.

Try Koji free

If you have been pasting customer transcripts into ChatGPT and wondering if there is a better way — there is. Start a free study with 10 credits at signup. Build a discussion guide, run an AI-moderated interview, and get a thematic report with verbatim-anchored quotes in 48 hours. No transcripts ever leave a research-grade environment, and every quote links back to the source recording.

For more, read how to analyze customer interview data, the GDPR and LLMs guide, or why AI interviewers are the future of customer research.

Make talking to users a habit, not a hurdle.