{"site":{"name":"Koji","description":"AI-native customer research platform that helps teams conduct, analyze, and synthesize customer interviews at scale.","url":"https://www.koji.so","contentTypes":["blog","documentation"],"lastUpdated":"2026-05-12T11:02:17.718Z"},"content":[{"type":"documentation","id":"829576cb-9021-4836-bd98-331df6303517","slug":"ai-research-agent","title":"AI Research Agent: How Autonomous AI Conducts User Interviews End-to-End","url":"https://www.koji.so/docs/ai-research-agent","summary":"An AI research agent is an autonomous, goal-directed system that handles the full user research loop — brief, interview, analysis, report — without a human moderator. Koji ships four cooperating agents (Consultant, Interviewer, Analyst, MCP server) that draft the brief, run voice or text interviews with adaptive probing across 6 structured question types, score transcript quality, and expose results to other agents (Claude, Cursor) via the Model Context Protocol. Unlike chatbots or AI-features-bolted-on survey tools, a true research agent makes interview-level decisions in real time and composes with other agents through machine-readable outputs.","content":"## The Bottom Line\n\nAn **AI research agent** is an autonomous system that handles the full user research loop — drafting the brief, interviewing participants conversationally with adaptive follow-ups, scoring response quality, and synthesizing themes — without a human moderator stepping in for any single conversation. Unlike a chatbot or a survey tool with \"AI features\" bolted on, a true research agent is goal-directed: you give it a research question and constraints, and it makes interview-level decisions in real time.\n\nKoji is built as an end-to-end agentic research platform. An **AI consultant** agent helps you scope the problem and draft the brief. An **AI interviewer** agent runs every conversation (voice or text), adapting probes per participant. An **AI analyst** scores each transcript, extracts themes, and aggregates per-question results. A **Model Context Protocol (MCP) server** then exposes everything to other agents (Claude, Cursor, your own scripts) as callable tools. The whole loop runs continuously without scheduling, transcription, or manual coding.\n\nThis article walks through what makes a research workflow truly \"agentic,\" what the four-agent stack inside Koji does at each step, and how AI research agents compare with traditional moderated interviews, survey tools, and AI-features-bolted-on platforms.\n\n## What \"AI Research Agent\" Actually Means\n\nThe term gets used loosely in 2026 marketing copy. To be precise, an AI research agent has four traits:\n\n1. **Goal-directed.** It works from a research objective, not a static script. If the participant says something interesting, the agent decides whether to probe or move on based on the brief's priorities.\n2. **Tool-using.** It can read context documents, query past studies, pull from a CRM, generate reports, or trigger webhooks — not just produce text.\n3. **Stateful.** It remembers what's been asked in this interview, which questions remain, and what quality bar each answer hit.\n4. **Composable with other agents.** Its outputs (transcripts, themes, structured answers) are machine-readable so other agents — like Claude or a coding agent in Cursor — can act on them.\n\nA single GPT prompt that \"writes survey questions\" is not an agent. A platform that lets an AI ask one canned question per page is not an agent. An agentic research system makes interview-level decisions autonomously and surfaces results in a form that other systems can consume.\n\n## The Four Agents Inside Koji\n\n### 1. The AI Consultant (Brief Generation Agent)\n\nWhen you start a new study, the AI consultant interviews **you** — the researcher. It clarifies the decision the research must inform, the current hypothesis, who should participate, and which methodology fits the question. The brief it produces is structured: a `problemStatement`, `decisionToInform`, `targetParticipant`, `methodology` framework (Mom Test, Jobs-to-be-Done, Customer Discovery, or custom), and an ordered list of `StudyQuestion` objects.\n\nUnlike a static template, the consultant pushes back. If you ask it to \"validate willingness to pay by asking would you pay $X,\" it will rewrite the question to anchor on past behavior — because the Mom Test framework explicitly forbids hypothetical pricing questions. The methodology is embedded as runtime principles, not a label.\n\n### 2. The AI Interviewer (Conversation Agent)\n\nEvery participant interview is run by the AI interviewer agent. Key behaviors:\n\n- **Mode-aware.** Voice or text, with a configurable probing depth per question (default: 1 follow-up; up to 3 for high-priority questions).\n- **Structured + exploratory blending.** Set the study to `structured` (cover every question in order), `exploratory` (follow interesting threads freely), or `hybrid` (default — cover the must-haves, free-roam on opportunities).\n- **Per-question follow-ups.** When a participant mentions a workaround, the agent asks how often it happens. When they say \"I wish…\", it asks what they've tried. These probe rules come from the methodology framework, not hardcoded.\n- **Six structured question types.** The agent natively understands `open_ended`, `scale`, `single_choice`, `multiple_choice`, `ranking`, and `yes_no` questions. It asks them conversationally (no \"click a radio button\" prompts in voice mode) but still extracts a chartable structured value.\n- **Multilingual.** Interviews run in 30+ languages, with the agent matching the participant's language even when the brief is in English.\n\n### 3. The AI Analyst (Transcript Synthesis Agent)\n\nThe moment an interview completes, the analyst agent runs:\n\n- Extracts a **quality score** (1–5) per interview, factoring in relevance, depth, coverage, completion rate, and structured-answer quality. Only conversations scoring 3 or higher consume credits — low-effort or abandoned sessions don't bill you.\n- Maps the conversation to each `StudyQuestion`, producing a `StructuredAnswer` (with `structuredValue`, `qualitativeAnswer`, and `confidence`) per question.\n- Tags emerging themes and aggregates them across all interviews in real time.\n- Flags surprising or anomalous findings for review.\n\n### 4. The MCP Server (Agent-to-Agent Bridge)\n\nThe outputs of all three agents above are exposed through a Model Context Protocol (MCP) server. That means **other agents** — Claude, Cursor, or any LLM with MCP support — can call Koji tools natively:\n\n- `koji_list_studies`, `koji_get_study`, `koji_get_interviews`, `koji_get_transcript`\n- `koji_create_study`, `koji_update_brief`, `koji_publish_study`\n- `koji_get_study_data`, `koji_generate_report`, `koji_publish_report`\n- `koji_configure_study`, `koji_export_data`, `koji_import_respondents`\n\nThis composability is what separates a research agent from a research SaaS with an AI chat feature. A product manager can say to Claude: *\"Pull the last 10 interviews from the onboarding study, group quotes by user persona, and draft a roadmap PRD.\"* Claude calls Koji's MCP tools, reads the data, and writes the doc — no manual export step.\n\n## How an AI Research Agent Differs From Adjacent Tools\n\n| Capability | Survey tools (Typeform, SurveyMonkey) | Recording-based research (UserTesting) | Repository-only AI (Dovetail) | **Koji AI research agent** |\n|---|---|---|---|---|\n| Drafts the research brief | No | No | No | Yes (AI consultant) |\n| Conducts the interview | No (static form) | Human-recorded session | No (analyzes existing recordings) | Yes (voice + text agent) |\n| Asks adaptive follow-ups | No | Only if a human moderator does | No | Yes (1–3 per question, autonomous) |\n| Scores transcript quality | No | Manual review | Partial | Yes (1–5 composite score) |\n| Aggregates themes in real time | No | After manual tagging | Yes | Yes |\n| Exposes results via MCP to other agents | No | No | No | Yes |\n| Time to first insight | Hours–days | 1–2 weeks | Depends on input pipeline | Minutes |\n\nThe gap to notice: every other category requires a human to either run the interview or feed the AI raw material. A true research agent removes both bottlenecks.\n\n## When to Use an AI Research Agent\n\n**Best fits:**\n\n- **Continuous discovery.** You want at least one customer interview per week without it eating your calendar. With tools like Koji, the AI interviewer runs 24/7 against one shared link.\n- **Cancel-flow and exit interviews.** Catch churning users mid-cancellation. By the time a human researcher could schedule a call, the user is gone. An agent interviews them in the moment.\n- **Onboarding friction studies.** Embed the agent in your in-product onboarding to interview new users about activation issues, with results auto-tagged by persona.\n- **Founder-led customer research.** A solo founder can run 50 customer-development interviews in a week — something that previously required hiring a researcher.\n- **B2B account research at scale.** Personalized interview links per account, with the agent referring to the company name and known pain points.\n- **Multilingual studies.** Run the same study in English, Spanish, German, and Japanese without four moderators.\n\n**Worse fits:**\n\n- **High-stakes legal/regulatory interviews** where every word needs human moderation.\n- **Co-design sessions** where the value is the collaborative whiteboard, not the transcript.\n- **Tiny ($N$ < 5) qualitative studies** where the time saved is marginal versus a 1:1 call.\n\n## Quality and Cost: How an AI Research Agent Stays Honest\n\nA frequent worry: \"If an AI runs the interview, how do I know it asked the right questions?\" Koji's agent is built to address this transparently.\n\n- **Brief-driven priorities.** Required questions must be covered before the agent ends the conversation. Optional questions are asked when there's time and context.\n- **Quality scoring per interview.** Every transcript gets a 1–5 score with a breakdown across relevance, depth, coverage, completion, and structured-answer quality. Interviews scoring 1 or 2 (rushed, off-topic, abandoned) **do not consume credits**.\n- **Credit-based pricing.** Text interviews cost 1 credit; voice interviews cost 3; report refreshes cost 5. The Insights plan includes 29 credits/month at €29; the Interviews plan includes 79 credits at €79. Overage is a flat €1/credit. No per-seat charges, no minimum commitments.\n- **Transparent transcripts.** Every interview transcript is viewable. Quotes in reports cite the source conversation. Exports (CSV/JSON) are available on all plans.\n\nThis matters because traditional moderated research has hidden quality variance too — a tired junior researcher running their tenth interview of the day isn't calibrated the same way as their first. An AI agent that scores its own transcripts and refunds credits for low-quality conversations is, in practice, more consistent than human moderation.\n\n## A Day in the Life: Running a Study With an AI Research Agent\n\n1. **9:00 AM** — You open Koji and tell the AI consultant: *\"We want to understand why trial users don't convert to paid.\"* The consultant clarifies your hypothesis (\"they don't see the value of the AI features\"), defines target participants (trial users in days 7–14, didn't convert), and proposes a Jobs-to-be-Done methodology. It drafts 8 questions across 2 sections.\n2. **9:15 AM** — You edit two questions, add a `scale` question (1–5: \"How likely are you to recommend us?\"), and publish the study.\n3. **9:20 AM** — You import 200 trial-user contacts from a CSV. The system generates personalized interview links for each.\n4. **9:30 AM** — You send an email blast with the links. Participants start clicking through.\n5. **10:00 AM** — The first 5 interviews complete. The dashboard already shows two emerging themes: \"AI features feel hidden\" and \"Free credits ran out before the user finished onboarding.\"\n6. **Throughout the week** — Interviews continue 24/7. The agent matches each participant's language, asks adaptive follow-ups, and routes transcripts to the analyst.\n7. **Friday** — You run `koji_generate_report` (or click \"Generate Report\"). The report cites direct customer quotes against each theme, shows the NPS distribution, and breaks down structured answers by persona segment.\n8. **Monday** — You hand the report URL to engineering. Or, you have Claude (via the MCP integration) read the report and draft a PRD for the team.\n\nThe whole study runs without a single human-moderated call.\n\n## Frequently Asked Questions\n\n**Are AI research agents biased?** They can be — exactly like human researchers. Koji mitigates this by embedding methodology principles directly in the agent prompt (Mom Test forbids leading questions; JTBD forbids product-feature questions) and by transparently exposing every transcript. You can see what the agent asked and how participants responded.\n\n**Can the agent fall for nonsense answers?** The quality score catches most of these — low-effort, contradictory, or off-topic responses score 1–2 and don't consume credits. You can manually flag any transcript for exclusion from the report.\n\n**What's the difference between Koji's agent and using ChatGPT for interviews?** ChatGPT is a general-purpose chatbot; it doesn't track which research questions have been covered, doesn't score conversation quality, doesn't aggregate across participants, and doesn't expose results as machine-readable tools. A research agent is a goal-directed system, not a single LLM call.\n\n## Related Resources\n\n- [AI Research Assistant: A Full Research Team in a Single Platform](/docs/ai-research-assistant) — broader overview of the agentic stack\n- [Understanding the AI Consultant](/docs/understanding-the-ai-consultant) — how the brief-generation agent works\n- [AI-Moderated Interviews: How Automated Research Works](/docs/ai-moderated-interviews) — deep dive on the interviewer agent\n- [Structured Questions Guide](/docs/structured-questions-guide) — the 6 question types the agent asks natively\n- [MCP Integration Overview](/docs/mcp-overview) — how other agents talk to Koji\n- [AI Interviewer Tuning Guide](/docs/ai-interviewer-tuning-guide) — getting research-grade voice interviews","category":"Reports & Analysis","lastModified":"2026-05-12T03:15:33.888034+00:00","metaTitle":"AI Research Agent: How Autonomous AI Runs User Interviews End-to-End","metaDescription":"AI research agents design studies, run conversational interviews, and synthesize findings without a moderator. See how Koji's four-agent stack handles the full research loop.","keywords":["ai research agent","autonomous user research","agentic research platform","ai research assistant","llm research agent","ai that runs user interviews","research agent koji"],"aiSummary":"An AI research agent is an autonomous, goal-directed system that handles the full user research loop — brief, interview, analysis, report — without a human moderator. Koji ships four cooperating agents (Consultant, Interviewer, Analyst, MCP server) that draft the brief, run voice or text interviews with adaptive probing across 6 structured question types, score transcript quality, and expose results to other agents (Claude, Cursor) via the Model Context Protocol. Unlike chatbots or AI-features-bolted-on survey tools, a true research agent makes interview-level decisions in real time and composes with other agents through machine-readable outputs.","aiPrerequisites":["Basic familiarity with user research workflows","A Koji account (free tier works)"],"aiLearningOutcomes":["Understand what makes a research workflow truly agentic","Map Koji's four-agent architecture (Consultant, Interviewer, Analyst, MCP)","Compare AI research agents against surveys, recording-based tools, and AI repositories","Decide when to use an AI research agent vs traditional moderated interviews","Understand quality scoring and credit-based pricing for autonomous interviews"],"aiDifficulty":"intermediate","aiEstimatedTime":"12 min read"}],"pagination":{"total":1,"returned":1,"offset":0}}