New

Now in Claude, ChatGPT, Cursor & more with our MCP server

Back to docs
Reports & Analysis

AI Research Agent: How Autonomous AI Conducts User Interviews End-to-End

A practical guide to AI research agents — autonomous systems that design studies, run conversational interviews, and synthesize findings without a human moderator. Includes how Koji's agentic stack works under the hood.

The Bottom Line

An AI research agent is an autonomous system that handles the full user research loop — drafting the brief, interviewing participants conversationally with adaptive follow-ups, scoring response quality, and synthesizing themes — without a human moderator stepping in for any single conversation. Unlike a chatbot or a survey tool with "AI features" bolted on, a true research agent is goal-directed: you give it a research question and constraints, and it makes interview-level decisions in real time.

Koji is built as an end-to-end agentic research platform. An AI consultant agent helps you scope the problem and draft the brief. An AI interviewer agent runs every conversation (voice or text), adapting probes per participant. An AI analyst scores each transcript, extracts themes, and aggregates per-question results. A Model Context Protocol (MCP) server then exposes everything to other agents (Claude, Cursor, your own scripts) as callable tools. The whole loop runs continuously without scheduling, transcription, or manual coding.

This article walks through what makes a research workflow truly "agentic," what the four-agent stack inside Koji does at each step, and how AI research agents compare with traditional moderated interviews, survey tools, and AI-features-bolted-on platforms.

What "AI Research Agent" Actually Means

The term gets used loosely in 2026 marketing copy. To be precise, an AI research agent has four traits:

  1. Goal-directed. It works from a research objective, not a static script. If the participant says something interesting, the agent decides whether to probe or move on based on the brief's priorities.
  2. Tool-using. It can read context documents, query past studies, pull from a CRM, generate reports, or trigger webhooks — not just produce text.
  3. Stateful. It remembers what's been asked in this interview, which questions remain, and what quality bar each answer hit.
  4. Composable with other agents. Its outputs (transcripts, themes, structured answers) are machine-readable so other agents — like Claude or a coding agent in Cursor — can act on them.

A single GPT prompt that "writes survey questions" is not an agent. A platform that lets an AI ask one canned question per page is not an agent. An agentic research system makes interview-level decisions autonomously and surfaces results in a form that other systems can consume.

The Four Agents Inside Koji

1. The AI Consultant (Brief Generation Agent)

When you start a new study, the AI consultant interviews you — the researcher. It clarifies the decision the research must inform, the current hypothesis, who should participate, and which methodology fits the question. The brief it produces is structured: a problemStatement, decisionToInform, targetParticipant, methodology framework (Mom Test, Jobs-to-be-Done, Customer Discovery, or custom), and an ordered list of StudyQuestion objects.

Unlike a static template, the consultant pushes back. If you ask it to "validate willingness to pay by asking would you pay $X," it will rewrite the question to anchor on past behavior — because the Mom Test framework explicitly forbids hypothetical pricing questions. The methodology is embedded as runtime principles, not a label.

2. The AI Interviewer (Conversation Agent)

Every participant interview is run by the AI interviewer agent. Key behaviors:

  • Mode-aware. Voice or text, with a configurable probing depth per question (default: 1 follow-up; up to 3 for high-priority questions).
  • Structured + exploratory blending. Set the study to structured (cover every question in order), exploratory (follow interesting threads freely), or hybrid (default — cover the must-haves, free-roam on opportunities).
  • Per-question follow-ups. When a participant mentions a workaround, the agent asks how often it happens. When they say "I wish…", it asks what they've tried. These probe rules come from the methodology framework, not hardcoded.
  • Six structured question types. The agent natively understands open_ended, scale, single_choice, multiple_choice, ranking, and yes_no questions. It asks them conversationally (no "click a radio button" prompts in voice mode) but still extracts a chartable structured value.
  • Multilingual. Interviews run in 30+ languages, with the agent matching the participant's language even when the brief is in English.

3. The AI Analyst (Transcript Synthesis Agent)

The moment an interview completes, the analyst agent runs:

  • Extracts a quality score (1–5) per interview, factoring in relevance, depth, coverage, completion rate, and structured-answer quality. Only conversations scoring 3 or higher consume credits — low-effort or abandoned sessions don't bill you.
  • Maps the conversation to each StudyQuestion, producing a StructuredAnswer (with structuredValue, qualitativeAnswer, and confidence) per question.
  • Tags emerging themes and aggregates them across all interviews in real time.
  • Flags surprising or anomalous findings for review.

4. The MCP Server (Agent-to-Agent Bridge)

The outputs of all three agents above are exposed through a Model Context Protocol (MCP) server. That means other agents — Claude, Cursor, or any LLM with MCP support — can call Koji tools natively:

  • koji_list_studies, koji_get_study, koji_get_interviews, koji_get_transcript
  • koji_create_study, koji_update_brief, koji_publish_study
  • koji_get_study_data, koji_generate_report, koji_publish_report
  • koji_configure_study, koji_export_data, koji_import_respondents

This composability is what separates a research agent from a research SaaS with an AI chat feature. A product manager can say to Claude: "Pull the last 10 interviews from the onboarding study, group quotes by user persona, and draft a roadmap PRD." Claude calls Koji's MCP tools, reads the data, and writes the doc — no manual export step.

How an AI Research Agent Differs From Adjacent Tools

CapabilitySurvey tools (Typeform, SurveyMonkey)Recording-based research (UserTesting)Repository-only AI (Dovetail)Koji AI research agent
Drafts the research briefNoNoNoYes (AI consultant)
Conducts the interviewNo (static form)Human-recorded sessionNo (analyzes existing recordings)Yes (voice + text agent)
Asks adaptive follow-upsNoOnly if a human moderator doesNoYes (1–3 per question, autonomous)
Scores transcript qualityNoManual reviewPartialYes (1–5 composite score)
Aggregates themes in real timeNoAfter manual taggingYesYes
Exposes results via MCP to other agentsNoNoNoYes
Time to first insightHours–days1–2 weeksDepends on input pipelineMinutes

The gap to notice: every other category requires a human to either run the interview or feed the AI raw material. A true research agent removes both bottlenecks.

When to Use an AI Research Agent

Best fits:

  • Continuous discovery. You want at least one customer interview per week without it eating your calendar. With tools like Koji, the AI interviewer runs 24/7 against one shared link.
  • Cancel-flow and exit interviews. Catch churning users mid-cancellation. By the time a human researcher could schedule a call, the user is gone. An agent interviews them in the moment.
  • Onboarding friction studies. Embed the agent in your in-product onboarding to interview new users about activation issues, with results auto-tagged by persona.
  • Founder-led customer research. A solo founder can run 50 customer-development interviews in a week — something that previously required hiring a researcher.
  • B2B account research at scale. Personalized interview links per account, with the agent referring to the company name and known pain points.
  • Multilingual studies. Run the same study in English, Spanish, German, and Japanese without four moderators.

Worse fits:

  • High-stakes legal/regulatory interviews where every word needs human moderation.
  • Co-design sessions where the value is the collaborative whiteboard, not the transcript.
  • Tiny ($N$ < 5) qualitative studies where the time saved is marginal versus a 1:1 call.

Quality and Cost: How an AI Research Agent Stays Honest

A frequent worry: "If an AI runs the interview, how do I know it asked the right questions?" Koji's agent is built to address this transparently.

  • Brief-driven priorities. Required questions must be covered before the agent ends the conversation. Optional questions are asked when there's time and context.
  • Quality scoring per interview. Every transcript gets a 1–5 score with a breakdown across relevance, depth, coverage, completion, and structured-answer quality. Interviews scoring 1 or 2 (rushed, off-topic, abandoned) do not consume credits.
  • Credit-based pricing. Text interviews cost 1 credit; voice interviews cost 3; report refreshes cost 5. The Insights plan includes 29 credits/month at €29; the Interviews plan includes 79 credits at €79. Overage is a flat €1/credit. No per-seat charges, no minimum commitments.
  • Transparent transcripts. Every interview transcript is viewable. Quotes in reports cite the source conversation. Exports (CSV/JSON) are available on all plans.

This matters because traditional moderated research has hidden quality variance too — a tired junior researcher running their tenth interview of the day isn't calibrated the same way as their first. An AI agent that scores its own transcripts and refunds credits for low-quality conversations is, in practice, more consistent than human moderation.

A Day in the Life: Running a Study With an AI Research Agent

  1. 9:00 AM — You open Koji and tell the AI consultant: "We want to understand why trial users don't convert to paid." The consultant clarifies your hypothesis ("they don't see the value of the AI features"), defines target participants (trial users in days 7–14, didn't convert), and proposes a Jobs-to-be-Done methodology. It drafts 8 questions across 2 sections.
  2. 9:15 AM — You edit two questions, add a scale question (1–5: "How likely are you to recommend us?"), and publish the study.
  3. 9:20 AM — You import 200 trial-user contacts from a CSV. The system generates personalized interview links for each.
  4. 9:30 AM — You send an email blast with the links. Participants start clicking through.
  5. 10:00 AM — The first 5 interviews complete. The dashboard already shows two emerging themes: "AI features feel hidden" and "Free credits ran out before the user finished onboarding."
  6. Throughout the week — Interviews continue 24/7. The agent matches each participant's language, asks adaptive follow-ups, and routes transcripts to the analyst.
  7. Friday — You run koji_generate_report (or click "Generate Report"). The report cites direct customer quotes against each theme, shows the NPS distribution, and breaks down structured answers by persona segment.
  8. Monday — You hand the report URL to engineering. Or, you have Claude (via the MCP integration) read the report and draft a PRD for the team.

The whole study runs without a single human-moderated call.

Frequently Asked Questions

Are AI research agents biased? They can be — exactly like human researchers. Koji mitigates this by embedding methodology principles directly in the agent prompt (Mom Test forbids leading questions; JTBD forbids product-feature questions) and by transparently exposing every transcript. You can see what the agent asked and how participants responded.

Can the agent fall for nonsense answers? The quality score catches most of these — low-effort, contradictory, or off-topic responses score 1–2 and don't consume credits. You can manually flag any transcript for exclusion from the report.

What's the difference between Koji's agent and using ChatGPT for interviews? ChatGPT is a general-purpose chatbot; it doesn't track which research questions have been covered, doesn't score conversation quality, doesn't aggregate across participants, and doesn't expose results as machine-readable tools. A research agent is a goal-directed system, not a single LLM call.

Related Resources