{"site":{"name":"Koji","description":"AI-native customer research platform that helps teams conduct, analyze, and synthesize customer interviews at scale.","url":"https://www.koji.so","contentTypes":["blog","documentation"],"lastUpdated":"2026-05-09T13:01:17.116Z"},"content":[{"type":"documentation","id":"1fd80e26-32b0-4c9b-8a3d-f36f13e2c9c4","slug":"open-axial-selective-coding","title":"Open, Axial, and Selective Coding: The Complete Guide to Qualitative Coding Phases","url":"https://www.koji.so/docs/open-axial-selective-coding","summary":"A complete pillar guide to the three coding phases of grounded theory: open coding (identifying concepts), axial coding (finding relationships using Strauss and Corbin's six-element coding paradigm), and selective coding (building a unifying theory around a core category). Includes worked examples, common mistakes, and a practical hybrid workflow combining AI-native research platforms like Koji with researcher judgment.","content":"## What are open, axial, and selective coding?\n\n**Open, axial, and selective coding are the three sequential phases of qualitative analysis in grounded theory — each one moves your data from raw text toward an explanatory theory.** Open coding breaks data into discrete concepts. Axial coding finds the relationships between those concepts. Selective coding chooses one core category and ties everything else to it.\n\nDeveloped by Anselm Strauss and Juliet Corbin in the 1990s as a refinement of Glaser and Strauss's original grounded theory ([Springer, 2019](https://link.springer.com/chapter/10.1007/978-3-030-15636-7_4)), the three-phase approach is now the most widely-used coding framework in qualitative research — taught in PhD programs, used by UX researchers, and applied in product discovery worldwide.\n\nThis guide walks through each phase with concrete examples, shows where teams typically get stuck, and explains how AI-native research platforms like **Koji** compress what used to be weeks of manual coding into automated thematic analysis you can validate and refine in hours.\n\n## Why grounded theory coding still matters\n\nIn an era of AI-assisted analysis, why bother understanding manual coding phases? Three reasons:\n\n1. **AI thematic analysis is a coding accelerator, not a replacement.** A 2025 tutorial in the *Journal of Medical Internet Research* found that ChatGPT-assisted coding \"enhanced the efficiency and diversity of coding\" but showed shortcomings in depth and context compared to manual work ([JMIR, 2025](https://www.jmir.org/2025/1/e70122)). Researchers who understand the underlying phases can validate and refine AI output; those who do not, cannot.\n\n2. **Grounded theory is how you build theory from data, not impose theory on data.** Open coding in particular is the discipline of staying open to what the data is telling you, rather than coding to confirm a hypothesis you already hold.\n\n3. **The three phases map onto how every product team thinks anyway** — observation → relationships → core insight. Knowing the formal vocabulary makes synthesis conversations sharper.\n\n## Phase 1: Open coding\n\n**Open coding is the first pass through your data, where you read line-by-line (or segment-by-segment) and label each meaningful unit with a short code.** The goal is breadth, not precision — capture every concept, action, or meaning you see, without pre-committing to categories.\n\n### How it works\n\n1. **Read or listen to your data once with no coding** — interview transcripts, field notes, open-ended survey responses. Get a feel for the whole.\n2. **Pass through again, line-by-line.** For each meaningful chunk, write a short label (a \"code\") that captures what is happening. Codes can be:\n   - **Descriptive** — \"first-time login confusion\"\n   - **In-vivo** — using participants' exact words: \"the dashboard is overwhelming\"\n   - **Process** — \"deciding-whether-to-upgrade\"\n3. **Keep codes provisional.** Resist the urge to consolidate yet. You will end up with hundreds of codes — that is normal and correct.\n4. **Write memos.** When a code feels significant or ambiguous, write a paragraph explaining what you saw. These memos become the connective tissue for axial coding.\n\n### Open coding example\n\nSuppose you ran six interviews about how product managers prepare for stakeholder reviews. A passage like:\n\n> \"I usually rebuild my deck from scratch the morning of, even though I have a template, because I never feel like the template covers what this particular VP cares about.\"\n\nMight produce open codes like:\n\n- last-minute deck rebuilding\n- template-not-fitting-audience\n- VP-specific preferences\n- avoiding-template-reuse\n- morning-of-review preparation\n\nNotice the codes are granular and overlapping. That is fine in open coding.\n\n### Common open coding mistakes\n\n- **Coding too sparsely.** If you have one code per page, you are summarizing, not coding. Aim for codes every 1–3 sentences.\n- **Importing pre-existing categories.** If you already \"know\" the themes, you will see only those themes. Open coding requires deliberate openness.\n- **Conflating codes too early.** \"Confusion\" and \"frustration\" feel similar but may have different antecedents. Keep them separate until axial coding tells you whether to merge.\n- **Skipping memos.** Memos are where the meaning lives. A spreadsheet of codes without memos loses the analytical reasoning behind each label.\n\n## Phase 2: Axial coding\n\n**Axial coding takes the codes from open coding and finds the relationships between them — what causes what, what conditions affect what, what consequences flow from what.** The output of axial coding is a smaller set of categories, each with subcategories that explain *how* the underlying codes connect.\n\n### Strauss and Corbin's coding paradigm\n\nThe canonical axial coding model uses a six-element framework — sometimes called the \"coding paradigm\" ([Springer, 2019](https://link.springer.com/chapter/10.1007/978-3-030-15636-7_4)):\n\n1. **Phenomenon** — the central concept being explained.\n2. **Causal conditions** — what gives rise to it.\n3. **Context** — the setting in which it occurs.\n4. **Intervening conditions** — broader structural factors that shape it.\n5. **Action/interaction strategies** — how people respond.\n6. **Consequences** — what happens as a result.\n\nNot every study uses all six elements. The framework is a scaffold, not a checklist.\n\n### Axial coding example\n\nUsing our open codes from earlier, axial coding might cluster them into:\n\n**Category: Audience-fit anxiety in stakeholder presentations**\n- *Causal conditions:* generic templates, varied stakeholder priorities\n- *Context:* high-stakes review meetings, limited prep time\n- *Intervening conditions:* lack of stakeholder profile data, no shared template library\n- *Action strategies:* last-minute rebuilding, asking peers for the latest VP version, padding decks with backup slides\n- *Consequences:* late nights, slide redundancy, decreased confidence\n\nNow you have a *category with structure*, not just a list of codes. A reader can see the causal logic.\n\n### When to move from open to axial\n\nMove when:\n- You stop seeing meaningfully new open codes (an early sign of saturation — see [Data Saturation in Qualitative Research](/docs/data-saturation-qualitative-research)).\n- Open codes are starting to cluster around recognizable concepts.\n- You have written enough memos to have a sense of the relationships at stake.\n\nIn practice, open and axial coding overlap. Most researchers iterate between the two before settling categories.\n\n## Phase 3: Selective coding\n\n**Selective coding is the final phase, where you choose one core category — the unifying concept that explains the most variance in your data — and integrate every other category around it.** This is where grounded theory earns its name: the output is a theory grounded in the data.\n\n### How to identify the core category\n\nStrauss and Corbin proposed several criteria. The core category should:\n\n- Appear frequently in the data\n- Connect to all (or most) other categories\n- Explain variation across cases\n- Be abstract enough to apply broadly, concrete enough to feel grounded\n- Have explanatory power — using it, can you make sense of why participants did what they did?\n\n### Selective coding example\n\nFrom the stakeholder presentation study, we might have several axial categories:\n\n- Audience-fit anxiety\n- Template-tool mismatch\n- Time-pressure cascading\n- Cross-team learning gaps\n\nThe selective coding question: is there one *core category* that ties these together?\n\nA candidate: **\"the gap between generic preparation tools and audience-specific expectations.\"** Each axial category becomes a facet of this core. Audience-fit anxiety is the felt experience; template-tool mismatch is the artifact-level cause; time-pressure cascading is the temporal consequence; cross-team learning gaps is the structural condition.\n\nNow you have a theory. You can write a research narrative around it. You can design product interventions against it.\n\n### When selective coding is too ambitious\n\nNot every qualitative study needs to produce a grounded theory. For most product and UX research, **axial coding output is sufficient**: a set of well-structured categories with relationships explained. Selective coding is the right move when:\n\n- You are doing dissertation-level or publication-targeted research.\n- The decision your insights inform is large enough to warrant a unifying explanation.\n- You have enough data (typically 20+ rich interviews) to support theoretical claims.\n\nFor a sprint-level study, stop at axial coding and skip selective. Forcing selective on thin data produces theories that overreach.\n\n## Open vs. axial vs. selective: at-a-glance\n\n| Phase | Goal | Output | Typical duration (manual) |\n|---|---|---|---|\n| Open | Identify concepts | 100s of codes + memos | Days to weeks |\n| Axial | Find relationships | 5–15 categories with subcategory structure | Days |\n| Selective | Build a unifying theory | One core category integrating all others | Days |\n\nManual grounded theory coding for a 20-interview study typically takes **80–120 hours of analyst time** across the three phases — which is why so few teams do it rigorously.\n\n## How AI-native research changes the workflow\n\nThis is where Koji enters the picture. Rather than spend 80–120 hours coding by hand, modern teams use AI to accelerate the heavy lift while preserving researcher judgment for the strategic moves.\n\n**With Koji:**\n\n- **AI-moderated interviews** generate clean transcripts automatically — no transcription cost, no waiting.\n- **Automatic thematic analysis** runs the equivalent of open + axial coding the moment interviews complete: clustering quotes, surfacing recurring themes, and labeling them in plain language.\n- **Quality scoring (1–5 scale)** identifies the richest interviews — the ones most likely to yield axial categories with explanatory depth.\n- **AI consultants** can be configured per-study to encode your research focus and coding priorities, so the AI's clustering reflects what *you* care about, not just what is statistically most frequent.\n- **6 structured question types** (open_ended, scale, single_choice, multiple_choice, ranking, yes_no) capture both the qualitative material that powers grounded coding and the quantitative anchors that ground theoretical claims. See the [Structured Questions Guide](/docs/structured-questions-guide).\n- **Real-time reports** let you see emerging categories as interviews land, so axial coding happens *during* fieldwork instead of after.\n\nThe critical principle: **AI accelerates open and axial coding; selective coding remains a human judgment call.** Koji surfaces candidate themes faster than any manual workflow can, but choosing the *core category* — the unifying explanation that earns the right to be called a theory — is editorial work that belongs to you.\n\nThe 2025 JMIR tutorial on ChatGPT for grounded theory captured the principle well: AI is strongest at the breadth-and-clustering end of coding, weakest at the depth-and-judgment end ([JMIR, 2025](https://www.jmir.org/2025/1/e70122)). Use Koji for the former; bring researcher expertise to the latter.\n\n## A practical workflow combining manual and AI coding\n\n1. **Run interviews via Koji** (AI-moderated voice or text — see [Setting Up Voice Interviews](/docs/setting-up-voice-interviews)).\n2. **Let Koji's automatic thematic analysis surface initial themes.** Treat these as candidate axial categories, not finished conclusions.\n3. **Read 3–5 full transcripts manually.** Even with AI clustering, deep reading is irreplaceable for catching nuance, contradictions, and emergent codes the AI missed.\n4. **Refine categories.** Merge AI-generated themes that overlap, split themes that conflate distinct concepts, add categories the AI missed.\n5. **Apply the coding paradigm to top categories.** For each major axial category, ask: what causes it, what consequences flow from it, what conditions shape it.\n6. **Identify the core category** (selective coding) — only if the study scope warrants it.\n7. **Write the research narrative** using the core category as the spine. See [Research Storytelling](/docs/research-storytelling).\n\n## Frequently asked questions\n\n**Do I need to do all three coding phases for every study?** No. Open and axial coding are sufficient for most product and UX research. Selective coding is appropriate for academic work or large strategic studies where a unifying theory is the deliverable.\n\n**What is the difference between open coding and thematic analysis?** Thematic analysis is a broader, more flexible methodology popularized by Braun and Clarke. It overlaps significantly with open + axial coding from grounded theory but is methodologically less prescriptive. See the [Thematic Analysis Guide](/docs/thematic-analysis-guide) for a full comparison.\n\n**How long does manual coding take?** Roughly 4–6 hours of coding per hour of interview, across all three phases. A 20-interview study (20 hours of audio) typically requires 80–120 hours of analyst time. AI-native platforms like Koji can compress the open/axial portion by 70–90%.\n\n**Can AI do grounded theory coding for me?** AI accelerates open and axial coding effectively, but selective coding requires human judgment. The 2025 JMIR tutorial found AI useful for breadth and clustering, weaker for depth and theoretical integration. Best practice is a hybrid workflow: AI for the heavy lift, human researcher for the strategic synthesis.\n\n**What is in-vivo coding?** In-vivo coding uses participants' exact words as codes — for example, \"the dashboard is overwhelming\" instead of \"interface complexity.\" It is a sub-technique within open coding that preserves participant voice and is especially valuable when terminology itself is part of the finding.\n\n**Do I need a software tool for grounded theory coding?** For studies under 5 interviews, a spreadsheet works. For 5–20 interviews, dedicated tools like NVivo, ATLAS.ti, or Delve speed manual coding. For ongoing or large-scale research, AI-native platforms like Koji handle the open/axial work automatically while you direct the strategic synthesis.\n\n## Related resources\n\n- [Structured Questions Guide](/docs/structured-questions-guide) — Combine open-ended and quantitative question types in interviews to power richer grounded theory analysis.\n- [Coding Qualitative Data](/docs/coding-qualitative-data) — Broader overview of qualitative coding approaches beyond grounded theory.\n- [Grounded Theory Qualitative Research](/docs/grounded-theory-qualitative-research) — The full methodology of which open/axial/selective coding is the analytical engine.\n- [Thematic Analysis Guide](/docs/thematic-analysis-guide) — A flexible alternative to grounded theory coding.\n- [Data Saturation in Qualitative Research](/docs/data-saturation-qualitative-research) — Knowing when to stop collecting and start coding.\n- [Research Synthesis Guide](/docs/research-synthesis-guide) — The broader synthesis workflow that selective coding feeds into.","category":"Analysis & Synthesis","lastModified":"2026-05-09T03:28:51.608194+00:00","metaTitle":"Open, Axial, Selective Coding Guide | Grounded Theory | Koji","metaDescription":"Master the three coding phases of grounded theory — open, axial, and selective coding. Step-by-step examples, common mistakes, and AI-native workflow with Koji.","keywords":["open coding","axial coding","selective coding","open coding qualitative research","grounded theory coding","qualitative coding phases","open vs axial coding","strauss and corbin coding"],"aiSummary":"A complete pillar guide to the three coding phases of grounded theory: open coding (identifying concepts), axial coding (finding relationships using Strauss and Corbin's six-element coding paradigm), and selective coding (building a unifying theory around a core category). Includes worked examples, common mistakes, and a practical hybrid workflow combining AI-native research platforms like Koji with researcher judgment.","aiPrerequisites":["Familiarity with qualitative research basics","At least one completed qualitative study","Understanding of what a research code is"],"aiLearningOutcomes":["Run open coding on raw qualitative data using line-by-line analysis and memos","Apply Strauss and Corbin's six-element coding paradigm to find relationships in axial coding","Identify a core category and build grounded theory through selective coding","Decide when selective coding is appropriate vs when to stop at axial","Combine AI-native platforms like Koji with manual coding for an accelerated hybrid workflow"],"aiDifficulty":"advanced","aiEstimatedTime":"15 minutes"}],"pagination":{"total":1,"returned":1,"offset":0}}