Knowledge Capture and Codification
There is a moment in every knowledge management initiative when someone says, "We need to get this out of people's heads and into a system." The impulse is understandable. People leave, retire, get sick, and forget. Systems — databases, documents, wikis — persist. The logic seems irrefutable: capture what experts know, write it down, and the organization becomes resilient against the inevitable departure of individuals.
The logic is also, in important ways, wrong. Or rather, it is right about the goal and dangerously simplistic about the method. The gap between what an expert knows and what can be captured in any external representation is not a minor inconvenience to be overcome with better templates and more thorough interviews. It is a fundamental feature of human knowledge, rooted in the nature of expertise itself.
This chapter examines the techniques for knowledge capture and codification — what works, what does not, and why the documentation paradox (everyone wants documentation; nobody wants to write it; nobody reads what gets written) is not a bug but a structural feature of how knowledge works.
The Documentation Paradox
Before discussing techniques, it is worth confronting the elephant in the room. Every organization complains about insufficient documentation. Every organization that invests in creating documentation discovers that much of it goes unread. And every effort to mandate documentation produces a predictable cycle: initial compliance, declining quality, eventual abandonment.
This is not because people are lazy or irresponsible (though some are). It is because documentation has a fundamental cost-benefit asymmetry. The cost of creating good documentation is borne by the author, now, in the form of time and cognitive effort. The benefit accrues to future readers, later, in unpredictable ways. The author rarely sees the benefit of their own documentation, and the future reader rarely appreciates the effort that went into creating it.
Moreover, documentation degrades. A perfectly accurate process document becomes misleading the moment the process changes. A troubleshooting guide becomes dangerous when it references components that have been replaced. The maintenance cost of documentation is ongoing and proportional to the rate of change in the domain it describes — which is precisely why documentation for fast-moving domains (software, technology, rapidly evolving business processes) is so often out of date.
The documentation paradox does not mean documentation is useless. It means that documentation strategies must be designed with the paradox in mind: minimize creation cost, maximize maintenance feasibility, target documentation at the areas where it provides the most value, and supplement it with other knowledge-transfer mechanisms.
Expert Interviews: Mining the Mother Lode
Expert interviews are the most direct method of knowledge capture. You sit down with someone who knows things and extract what they know. The technique sounds simple. It is not.
Structured Interviews
A structured knowledge capture interview differs from a journalistic interview or a casual conversation. It follows a predefined protocol designed to elicit specific types of knowledge:
Process knowledge: "Walk me through how you do X." The interviewer asks the expert to describe their workflow step by step, probing at each step for decision points, exceptions, and alternatives. The goal is to surface not just the standard procedure but the variations, shortcuts, and judgment calls that distinguish expert performance from competent-but-routine performance.
Decision knowledge: "How do you decide between X and Y?" Decision-focused questioning aims to elicit the criteria, heuristics, and mental models that experts use when making choices. Critical Decision Method (CDM), developed by Gary Klein and colleagues, uses a specific protocol: identify a challenging incident, construct a timeline, probe for decision points, and explore what information the expert used, what alternatives they considered, and what cues triggered their choices.
Troubleshooting knowledge: "What do you do when things go wrong?" Experts often possess extensive knowledge about failure modes, diagnostic strategies, and recovery procedures that is poorly documented because it deals with situations that are not supposed to happen.
Contextual knowledge: "What do you need to know about the environment to do this well?" Experts have knowledge about the context in which their work occurs — organizational politics, supplier relationships, seasonal variations, historical reasons for current practices — that is essential for effective performance but rarely documented.
The Problem of Tacit Knowledge
The fundamental challenge of expert interviewing is that experts often cannot articulate what they know. This is not false modesty or deliberate concealment. It is a well-documented feature of expertise, explored by Michael Polanyi in The Tacit Dimension (1966) and confirmed by decades of subsequent research.
Expert knowledge is, to a significant degree, tacit — embedded in perceptual skills, motor routines, pattern recognition capabilities, and intuitions that operate below the level of conscious awareness. A master chess player does not calculate every possible move; they perceive the board in terms of patterns and opportunities that they could not fully describe. An experienced physician does not diagnose by running through a mental checklist of symptoms; they recognize patterns in ways that feel more like perception than reasoning.
This creates an inherent limit on knowledge capture. No matter how skilled the interviewer or how willing the expert, some knowledge will resist articulation. The practical implication is that knowledge capture should be supplemented by other mechanisms — apprenticeship, observation, simulation, worked examples — that transmit tacit knowledge through practice rather than through language.
Practical Guidelines for Knowledge Capture Interviews
If you are conducting knowledge capture interviews, several practices improve the yield:
Use concrete examples, not abstractions. Asking "How do you handle difficult customers?" will produce vague generalities. Asking "Tell me about a specific time when a customer situation was particularly challenging" will produce detailed, actionable narratives. Experts reason from cases, not from principles.
Probe for exceptions and edge cases. The standard procedure is the easy part. The valuable knowledge is in the exceptions: "What do you do when this does not work?" "When would you deviate from the standard approach?" "What is the weirdest case you have ever encountered?"
Watch them work, then ask about it. Observation-based interviews (sometimes called contextual inquiry) combine observation with questioning. You watch the expert perform a task, note what they do, and then ask why. This surfaces knowledge that the expert would not think to mention in a purely verbal interview because it is so automatic that they do not notice it.
Record and transcribe. Real-time note-taking during an interview inevitably loses nuance and detail. Audio recording (with the expert's consent) preserves the full conversation for later review. Video recording adds the ability to capture gestures, demonstrations, and interactions with tools and artifacts.
Iterate. A single interview rarely captures everything. Multiple sessions, with time between them for the interviewer to review and identify gaps, produce far better results.
After-Action Reviews
The after-action review (AAR) is a structured method for learning from experience, originally developed by the U.S. Army in the 1970s and subsequently adopted by organizations ranging from hospitals to software companies.
The format is deceptively simple. After a project, event, or significant activity, participants gather to answer four questions:
- What was supposed to happen? (The plan, the expected outcome.)
- What actually happened? (The facts, without blame or interpretation.)
- Why was there a difference? (Root cause analysis.)
- What can we learn from this? (Actionable lessons for the future.)
The power of the AAR lies not in the questions — which are obvious — but in the discipline of actually conducting it and the norms that govern the discussion. Effective AARs require:
Timeliness: Conduct the AAR as soon as possible after the event, while memories are fresh. An AAR conducted six months after a project is an exercise in collective confabulation.
Psychological safety: Participants must be able to speak honestly about what went wrong without fear of punishment. This is why the Army's AAR protocol emphasizes that rank is suspended during the review — a norm that many corporate cultures struggle to replicate.
Focus on systemic causes, not individual blame: "Why did the process fail?" rather than "Who screwed up?" Blame produces defensiveness and silence. Systemic analysis produces actionable improvements.
Documentation and follow-through: An AAR that produces lessons but no changes is worse than useless — it teaches people that the organization does not actually learn, discouraging future participation. Lessons must be assigned owners, tracked, and implemented.
The British Army uses a similar process called a "hot debrief" (immediately after an event) and "cold debrief" (after a longer interval, allowing more reflective analysis). Some organizations distinguish between "hot" AARs (quick, tactical, focused on immediate improvements) and "cold" AARs (thorough, strategic, focused on systemic patterns).
Lessons Learned Databases
A lessons learned database is a repository where the outputs of AARs, project reviews, and other reflective processes are stored for future reference. In theory, this is a powerful tool: instead of each team learning from its own mistakes, the entire organization can learn from everyone's mistakes.
In practice, lessons learned databases have a dismal track record. NASA's Lessons Learned Information System, discussed in the previous chapter, is the canonical cautionary tale, but the problems are widespread. Common failure modes include:
Low contribution rates: People do not submit lessons because the process is cumbersome, because they do not see evidence that submitted lessons are read, or because the organizational culture does not reward the vulnerability required to admit mistakes.
Low retrieval rates: People do not search the database because they do not know it exists, because the search is poor, because they do not think to look (people starting a new project are focused on the future, not the past), or because the lessons are not organized in a way that maps to their current situation.
Quality degradation: Lessons are often written at too high a level of abstraction to be useful ("Communication is important" — thank you, very helpful) or at too low a level of specificity to be transferable ("We should have ordered the titanium fasteners from Supplier X rather than Supplier Y" — not useful if you are not building the same thing).
Staleness: Lessons from ten years ago may be irrelevant or actively misleading in a changed context. Few organizations invest in curating and retiring old lessons.
The organizations that make lessons learned databases work tend to share several characteristics: they integrate lesson retrieval into existing workflows (rather than requiring people to separately search the database), they use structured formats that make lessons scannable and filterable, they assign accountability for lesson follow-through, and they periodically review and cull the database to maintain quality.
AI offers a potentially transformative improvement here. Instead of requiring people to search a lessons learned database, a RAG-enabled system can proactively surface relevant lessons when it detects that a team is working on a problem similar to one that has been encountered before. This shifts the burden from the knowledge seeker (who may not know what to search for) to the system (which can monitor context and make suggestions).
Decision Logs
A decision log is a record of significant decisions: what was decided, when, by whom, what alternatives were considered, what information was available, and what reasoning led to the chosen course of action. Decision logs serve two purposes: they enable future decision-makers to understand why things are the way they are (avoiding the "why on earth did they do it this way?" problem), and they provide material for post-hoc analysis of decision quality.
Effective decision logs capture:
- The decision itself: What was decided, stated clearly and unambiguously.
- Context: What was the situation that prompted the decision? What constraints were in play?
- Alternatives considered: What other options were on the table? Why were they rejected?
- Reasoning: What logic, evidence, or judgment led to the chosen option?
- Expected outcomes: What did the decision-makers expect to happen as a result?
- Decision-makers: Who was involved? Who had the final say?
- Date: When the decision was made (critical for understanding what information was available at the time).
The most common failure of decision logs is that they record the what without the why. A log entry that says "Decided to use PostgreSQL for the new application" is almost useless. An entry that says "Decided to use PostgreSQL because we need strong transactional integrity for financial data, the team has deep PostgreSQL expertise, and Oracle licensing costs were prohibitive; considered MongoDB (rejected due to consistency requirements) and MySQL (rejected due to limited JSON support at the time)" is genuinely valuable.
Architectural Decision Records (ADRs), popularized by Michael Nygard in 2011, are a lightweight format for decision logging in software development. Each ADR is a short document with a fixed structure — Title, Status, Context, Decision, Consequences — stored alongside the code it relates to. The format has been widely adopted because it is low-overhead and directly integrated into the development workflow.
Process Documentation
Process documentation describes how work is done: the steps, the inputs and outputs, the roles and responsibilities, the decision points, and the exception handling. It is the most common form of knowledge codification and, for routine work, the most valuable.
Good process documentation has several characteristics:
It is task-oriented: Written from the perspective of someone trying to do the work, not from the perspective of someone trying to describe the system. "To process a refund, first verify the original transaction in the billing system" rather than "The refund subsystem interfaces with the billing module via the transaction verification API."
It is layered: Different audiences need different levels of detail. A high-level process overview serves managers and new employees; a detailed step-by-step procedure serves practitioners; a technical reference serves system administrators. These layers should be linked, not collapsed into a single document.
It is maintained: This is the hard part. Process documentation must be updated when the process changes, and the responsibility for updating it must be clearly assigned. The most sustainable approaches tie documentation updates to process changes: the same pull request that changes the code updates the documentation; the same process improvement initiative that modifies a workflow updates the process document.
It distinguishes between normative and descriptive: There is a difference between how work is supposed to be done (normative documentation — the official procedure) and how work is actually done (descriptive documentation — the reality). In many organizations, these diverge significantly, and the official documentation describes a process that nobody actually follows. Useful process documentation acknowledges this gap and either updates the documentation to match reality or updates the process to match the documentation.
Knowledge Representation
When knowledge is captured, it must be represented in some structured form if it is to be stored, searched, and reasoned about. Several formal knowledge representation schemes have been developed, primarily in the artificial intelligence and cognitive science communities.
Frames (developed by Marvin Minsky in the 1970s) represent knowledge as structured records with named slots and default values. A "restaurant" frame might have slots for cuisine type, price range, location, hours, and typical-meal-sequence, with defaults that can be overridden for specific instances. Frames capture the structured, expectation-based nature of much human knowledge.
Scripts (developed by Roger Schank and Robert Abelson in the 1970s) represent knowledge about stereotypical event sequences. A "restaurant script" describes the expected sequence of events when dining at a restaurant: enter, be seated, order, eat, pay, leave. Scripts capture procedural knowledge and enable inference about events that are not explicitly mentioned (if someone describes eating at a restaurant and then leaving, you can infer that they paid).
Semantic networks represent knowledge as graphs of nodes (concepts) and edges (relationships). "A canary is a bird," "a bird can fly," "a canary is yellow" form a simple semantic network that supports inheritance-based reasoning (since a canary is a bird and birds can fly, a canary can fly). Semantic networks are the ancestors of modern knowledge graphs and ontologies.
Production rules represent knowledge as condition-action pairs: "IF the patient has a fever AND a rash, THEN consider measles." Expert systems of the 1980s were built primarily on production rules, and the format remains useful for representing diagnostic and decision-making knowledge.
Concept maps and mind maps are less formal but widely used representations. Concept maps (developed by Joseph Novak in the 1970s) show concepts connected by labeled relationships, while mind maps (popularized by Tony Buzan) show ideas radiating from a central theme. Both are useful for knowledge elicitation and communication, though they lack the formal semantics needed for computational reasoning.
These representation schemes are not merely academic curiosities. When you design a knowledge base, you are implicitly choosing a representation scheme. A wiki page is an informal frame. A troubleshooting flowchart is a set of production rules. A tag structure is a rudimentary semantic network. Understanding the formal schemes helps you design better informal ones, because you understand what each representation can and cannot express.
The Gap Between Knowing and Telling
The most important thing to understand about knowledge capture is that it will always be incomplete. This is not a counsel of despair — it is a design constraint. If you design your knowledge management system on the assumption that all relevant knowledge can be captured and codified, you will be disappointed. If you design it on the assumption that codification is one mechanism among several, complemented by communities of practice, mentoring, apprenticeship, and other person-to-person transfer mechanisms, you will be more successful.
The gap between knowing and telling has several dimensions:
Experts do not know what they know. Much expert knowledge is compiled — automated through practice to the point where it operates below conscious awareness. Ask an expert driver how they parallel park, and they will give you a simplified, post-hoc rationalization that omits most of what they actually do. The cognitive science literature calls this the "curse of expertise": the more expert you are, the harder it is to articulate your knowledge, because you have forgotten what it is like not to know.
Context is everything. Knowledge that is perfectly clear in context becomes ambiguous or meaningless out of context. A troubleshooting tip that makes perfect sense to someone familiar with the system may be gibberish to a newcomer. Capturing the context — the assumptions, the background knowledge, the physical and organizational environment — is often harder than capturing the knowledge itself.
Knowledge is dynamic. What an expert knows today is not what they knew yesterday or will know tomorrow. Knowledge evolves through experience, and a knowledge capture exercise is a snapshot, not a continuous recording. The snapshot begins degrading the moment it is taken.
Some knowledge is embodied. Riding a bicycle, performing surgery, throwing a pot on a wheel — these involve knowledge that is inseparable from the physical skills involved. No amount of documentation will transfer this kind of knowledge. It requires practice, feedback, and often the physical presence of an experienced practitioner.
The practical response to these limitations is not to abandon codification but to be strategic about it. Codify what can be codified — procedures, decision criteria, factual information, organizational context. For the rest, invest in the social and experiential mechanisms that transfer tacit knowledge: apprenticeship, pairing, communities of practice, storytelling, simulation, and hands-on exercises.
And when you do codify knowledge, include enough context that future readers can assess whether the knowledge is still applicable. A decision log entry that includes the reasoning and the context allows future readers to judge whether the decision still makes sense in changed circumstances. A lessons learned entry that includes the situation and the constraints allows future readers to assess whether the lesson transfers to their situation. Context is not optional metadata — it is what makes captured knowledge usable.
Knowledge Capture in the Age of AI
Large language models and related AI technologies are changing the knowledge capture landscape in several ways.
Automated summarization can reduce the effort required to produce documentation from meetings, interviews, and discussions. Instead of someone taking notes and writing them up, an AI system can produce a first draft from a recording. This does not eliminate the need for human review and editing — AI summaries can miss nuance, misrepresent emphasis, and hallucinate details — but it significantly reduces the activation energy for documentation.
Conversational elicitation is an emerging technique where an AI system interviews an expert, asking follow-up questions, probing for details, and organizing the resulting knowledge into structured formats. The AI can be tireless, thorough, and systematic in ways that human interviewers sometimes are not. Early implementations are promising, though the AI's inability to truly understand the domain limits its ability to probe deeply.
Continuous capture becomes more feasible when AI can process unstructured data — emails, chat messages, meeting recordings, code commits — and extract knowledge artifacts. Instead of requiring explicit knowledge capture activities, the system captures knowledge as a byproduct of normal work. This is attractive in theory but raises significant concerns about privacy, consent, and the quality of knowledge extracted from informal communications.
Knowledge graph construction can be partially automated using AI. Named entity recognition, relationship extraction, and ontology learning can identify concepts and relationships in unstructured text and populate a knowledge graph. This does not replace human curation — automated extraction produces noisy, incomplete graphs — but it provides a starting point that is far less expensive than manual construction.
The fundamental dynamics of knowledge capture — the documentation paradox, the tacit knowledge gap, the maintenance burden — are not eliminated by AI. But the cost-benefit equation shifts. When the cost of creating documentation drops dramatically, it becomes feasible to document things that were previously not worth the effort. When AI can assist with maintenance — flagging outdated content, suggesting updates, merging duplicates — the maintenance burden becomes more manageable. The result is not a solution to the knowledge capture problem but a significant improvement in the terms on which it is managed.