Deadwater.ai

apr 06 2026

Why every serious AI team is building a context layer

AGENTS.md and CLAUDE.md are not the story. They are early signs that companies are starting to build context as infrastructure.

15 min read
context-engineeringai-agentscontent-osmarkdownknowledge-systems
Why every serious AI team is building a context layer

Why every serious AI team is building a context layer

The next layer in software is not another dashboard. It is company context, turned into something machines can actually use.

For a while, AI adoption looked like a prompt problem.

People swapped prompts in Slack, saved favorite snippets in Notion, and kept pretending the next clever instruction block would finally make the model "get it." Then reality did what reality usually does. It ruined the fantasy. The prompt worked on Tuesday, failed on Thursday, ignored product constraints on Friday, and by the following week somebody was manually pasting the same company background into five different tools like a deranged intern.

That phase is not exactly over, but it is getting exposed. You can see it in the tools now. OpenAI's Codex launch explicitly points to AGENTS.md as a way to guide repository-level behavior. Anthropic's Claude Code memory docs formalize CLAUDE.md as a project memory surface. Anthropic also said the quiet part out loud in its piece on effective context engineering: the real work is no longer just model prompting. It is building the right context around the model.

That shift matters because AGENTS.md is not the revolution, and CLAUDE.md is not the revolution either. They are the first visible artifacts of a much bigger transition: companies are starting to externalize their knowledge, rules, operating constraints, and workflows into a machine-usable context layer.

That is where this is going. Not toward one giant prompt, infinite context, or your AI becoming an omnipotent employee because you dumped the company drive into a vector store and prayed. Toward a context stack, and toward a world where the teams that build that stack well pull away from the teams still doing prompt cosplay.

Prompts are giving way to context systems

The problem was never just wording

If you have used AI seriously inside a business, you already know this.

The model usually does not fail because your prompt was missing one magical sentence. It fails because it does not know enough about your business, your constraints, your definitions, your customers, your standards, or the shape of the task it is being asked to complete.

That is why the "prompt engineering" era started feeling weirdly flimsy the moment people tried to operationalize it.

A prompt can tell the model to "act like a senior product marketer." Fine. Very cinematic. But can it tell the model:

  • Which product claims are safe to make.
  • Which audience objections matter right now.
  • Which naming conventions changed last month.
  • Which workflows require review before publishing.
  • Which internal definitions should override the model's prior assumptions.

That is a different category of problem.

You are not asking for prettier wording anymore. You are asking for grounded behavior.

This is where a lot of AI discussions still go soft. They talk about intelligence when the real issue is context exposure. They talk about creativity when the real issue is reliable access to source truth. They talk about bigger models when the actual failure mode is that the system is still running on scraps, guesses, and copy-pasted background information.

That is why the useful shift is from prompt engineering to context strategy. The question is no longer "what should I type into the box?" It is "what should this system know before it acts?"

The market is already moving there

The official docs are only one signal, but they are a strong one.

When OpenAI tells users to place repository instructions in AGENTS.md, that is a sign the operating surface is getting standardized. When Anthropic gives CLAUDE.md a defined place in the memory hierarchy, that is another signal. These are not random UX quirks. They reflect the same basic reality: if the system is going to operate repeatedly inside a codebase or workflow, it needs stable memory and stable rules.

You can see the same pattern outside coding tools.

Google Cloud's documentation for knowledge documents stresses relevance, freshness, and structure. That is not just retrieval hygiene. It is a warning against the corporate instinct to throw everything into the machine and call it a knowledge base. The same principle shows up across agent docs, RAG docs, and every halfway serious conversation about AI operations.

More context is not automatically better. Better context is better, and that distinction is going to matter more than most teams think.

Why this is happening now

Because the task changed. Early AI usage was mostly toy work, one-off assistance, novelty content, and lightweight synthesis. In that world, a smart model with a half-decent prompt looked impressive enough. Once teams started asking AI to produce production-grade work, update live systems, draft pages, review code, write structured documents, or operate inside repeatable workflows, the weaknesses got obvious fast.

A model cannot safely operate on vibes. It needs a map. That map is the beginning of a Content OS, whether people use that term yet or not. The moment you start defining what the system should know, what it should do, what rules it should follow, and what sources count as truth, you are no longer just "using AI." You are designing an operating layer around it. That is the bigger story.

What the new context layer actually looks like

It is not one file

This is where a lot of people are going to get lazy.

They see AGENTS.md or CLAUDE.md and assume the answer is to write one giant instruction file full of preferences, random lore, and overexplained process notes. Then six weeks later the file is bloated, stale, internally contradictory, and expensive to load.

Congratulations. You built a context landfill.

The useful version is smaller and more layered.

A real context layer usually includes things like:

  • Operating instructions for the local environment.
  • Product briefs and offer definitions.
  • Writing style guides and brand rules.
  • Workflow contracts and publishing requirements.
  • Structured docs with ownership, status, and update history.
  • Internal pages, docs, and modules that act as source truth.

Some of that may live in markdown files. Some may live in schemas. Some may live in code or metadata. The point is not the file extension. The point is that the system now has a readable, maintainable surface where company truth becomes available to machines.

That is a major architectural change disguised as documentation.

The best versions separate layers on purpose

One of the easiest mistakes here is mixing every kind of knowledge together.

Do not put stable business definitions, volatile status updates, temporary notes, editorial rules, and execution permissions into one giant blob. That just turns retrieval into mud.

The better pattern is layered context.

For example:

context_stack:
  load_order:
    - AGENTS.md
    - CLAUDE.md
    - workflow-policies.md
  business_truth:
    - product-brief.md
    - pricing-notes.md
    - positioning.md
  execution_rules:
    - style-guide.md
    - publishing-checklist.md
    - qa-rules.md
  knowledge_modules:
    - docs/
    - comparisons/
    - decision-records/
  safeguards:
    review_required_for:
      - homepage-copy
      - pricing-changes
      - legal-claims
    stale_after_days: 90

This is not about making things look neat for engineers with a folder fetish.

It changes behavior.

When context is layered, the system can reason about what kind of truth it is using. It can distinguish stable rules from current facts. It can load the right material for the right task. It can avoid dragging half the company's intellectual junk drawer into every interaction.

That is why markdown knowledge systems and living docs for agents matter. They are not cute documentation ideas. They are design patterns for machine-usable memory.

If you want the shorter version, it looks something like this in practice:

request: "Draft a new comparison page for Product X"
agent_reads:
  - AGENTS.md
  - product-brief.md
  - positioning.md
  - style-guide.md
  - comparisons/product-x.md
agent_checks:
  - claim-support-required
  - internal-link-opportunities
  - review-required-if-pricing-mentioned
agent_output:
  status: review
  file: /content/read/product-x-alternatives.md

Interactive layer map

What the context stack is actually made of

Click through the layers. The point is not to create one giant memory blob. The point is to give the agent a clean operating stack with different kinds of truth in the right places.

Selected layer

Instruction layer

Tells the agent how to behave in this environment before it starts improvising.

Usually includes

  • AGENTS.md
  • CLAUDE.md
  • Local repo rules

If this is weak

The model behaves generically and starts guessing how your system works.

Operator takeaway

A context layer is not one file. It is a stack of instruction, truth, execution, knowledge, and control surfaces that let the agent act without freewheeling.

This is how AI starts becoming operational

A lot of companies still have AI trapped in assistant mode.

Ask a question. Get a response. Maybe copy it into another tool. Maybe ask for a rewrite. Maybe manually paste more context in because the first answer was generic sludge.

That workflow does not scale. It also does not compound.

A context layer is what turns ad hoc assistance into repeatable operation.

Now the system can work from a known base of instructions, definitions, and source truth. Now it can draft with the current offer in mind. Now it can review against an actual style guide. Now it can decide whether something requires human approval because the authority model is written down. Now it can follow process without somebody standing over it like a camp counselor.

That does not mean the AI becomes magical. It means the environment becomes legible.

There is a big difference.

If you want machines to do serious work, you have to stop treating business context like background noise and start treating it like infrastructure.

Why markdown keeps winning as the control surface

Markdown is boring in exactly the right way

There is a reason this context layer keeps collapsing toward markdown and adjacent plain-text formats.

Markdown is easy to read, easy to diff, easy to version, easy to move, easy to audit, and easy to edit without begging some proprietary interface for permission. It is friendly to humans and legible to machines. That combination is absurdly useful.

This is also why the repo is becoming such an important operating surface for AI work. Git already gives you version control, history, rollback, and collaboration. Markdown gives you a durable way to express instructions, system notes, runbooks, decisions, and source-of-truth modules inside that environment.

That is a powerful combo.

And yes, there is some irony here. After years of software trying to hide complexity behind polished SaaS interfaces, the practical answer for AI operations keeps drifting back toward text files in a repo.

Sorry, but it makes sense.

Portability is becoming a strategic advantage

One underrated reason markdown matters is that it is not tied to a single vendor's worldview.

If your company context only exists inside one tool's proprietary memory layer, you are renting your operating knowledge from that tool. If the tool changes, the feature disappears, the pricing shifts, or the workflow needs to move somewhere else, you are stuck dragging your business context out of a black box.

That is a stupid dependency to create on purpose.

A markdown-based context layer is portable. It can travel between tools, support different agents, and plug into different runtimes. OpenAI can read one surface. Anthropic can read another. Internal tooling can read the same files. Humans can inspect them without a decoder ring. That matters more as the tool market gets noisier.

This is one of the reasons headless CMS architecture is not enough on its own. Storage and delivery are useful. But if the context that governs behavior is trapped inside brittle interfaces or invisible tool state, your AI stack is still fragile.

Portable context is stronger than clever tooling.

Plain text does not mean unstructured

This is where people get confused.

Markdown is not useful because it is loose. It is useful because it can hold structure without becoming miserable to maintain.

You can add frontmatter. You can define heading contracts. You can embed tables, checklists, version notes, ownership fields, review cycles, and machine-readable metadata. You can pair that with JSON Schema or validation logic when the workflow needs stricter guarantees.

That means you get a control surface that supports both interpretation and enforcement.

A file can say:

  • Who owns this.
  • What sections are mutable.
  • When it was last verified.
  • Which rules apply.
  • Which workflow it belongs to.
  • What kind of object it is.

That is not just documentation anymore. That is operational state.

Once teams realize that, they stop asking "should we make a knowledge base?" and start asking better questions:

  • Which documents are source truth?
  • Which rules need to be machine-readable?
  • Which context should load by default?
  • Which context should stay out unless explicitly needed?
  • How do we keep this current without turning it into sludge?

Those are the right questions.

Build this on a real Content OS

This post is one piece of the system. See how Deadwater structures content so AI can operate on it safely and at scale.

More context is not the win

"Infinite context" is mostly a coping mechanism

There is a certain kind of AI take that sounds smart until you have to actually operate it.

The basic version goes like this: models are getting larger context windows, retrieval is getting better, memory is getting richer, therefore the answer is to give the system as much data as possible and let it figure things out.

This sounds sophisticated right up until you watch the outputs get slower, noisier, more expensive, and less trustworthy.

Big context windows are useful. They are not a substitute for judgment.

Anthropic's own context engineering guidance is basically a polite version of this point. Same with Google Cloud's warnings about irrelevant or outdated knowledge documents. Same with every team that has spent a month cleaning up a bloated internal AI setup after realizing the assistant kept pulling from stale junk and low-signal documents.

The problem is not scarcity anymore. It is filtration.

Noise scales too

This is the part people miss when they fantasize about the AI knowing "everything."

Everything includes contradictions.

Everything includes stale pricing decks, abandoned strategy docs, old process notes, dead product names, random brainstorms, obsolete instructions, duplicated pages, and all the other sediment every company accumulates over time.

If you expose all of that without curation, the system does not become omniscient. It becomes politically confused.

That is why what belongs in an AI knowledge base for marketing teams makes the stricter argument it does. The issue is not access alone. It is curation, ownership, shape, and maintenance.

The winning teams are not the ones feeding the machine the biggest pile. They are the ones designing the cleanest path through the pile.

Reliable context has boundaries

This is where governance for agents enters the picture.

A serious context layer is not just a memory layer. It is also a control layer.

It defines what the system can trust, what it can change, what it can propose, what it can publish, and what still requires review. Without those boundaries, "context" becomes a weird mix of knowledge and authority, and the workflow starts drifting into chaos.

A practical context layer needs to answer at least four things:

  • What is true.
  • What is current.
  • What is allowed.
  • Who is responsible.

That is not glamorous. It is also where most of the real leverage lives.

Because once those boundaries exist, AI can move much faster without becoming reckless. The system knows where the floor is. Humans know where to intervene. The work stops depending on memory, vibes, and one overworked operator translating the business into prompts all day.

That is what operational maturity looks like.

What companies should build now

Start smaller than your instincts want

Most teams should not begin by trying to encode the entire company into machine-readable form. That way lies rot. Start with the context that directly improves high-value workflows:

  • Product and offer truth.
  • Audience and positioning truth.
  • Style and publishing rules.
  • Operational instructions.
  • Review and QA requirements.

That alone is enough to materially improve drafting, editing, coding, planning, and update workflows. It also gives you a manageable surface to maintain. The move is not "capture everything." It is "identify the truth that the system keeps needing, then expose that truth cleanly."

Build a stack, not a blob

If you are doing this seriously, the architecture should feel more like a stack than a scrap pile.

A useful version usually looks something like this:

  1. A thin instruction layer for tool behavior and environment-specific rules.
  2. A business-truth layer for product, positioning, pricing, audience, and constraints.
  3. An execution layer for style, QA, workflow contracts, and publishing rules.
  4. A knowledge layer for reusable docs, modules, and internal references.
  5. A governance layer for authority, review, and change tracking.

Not every team needs all five layers fully formalized on day one.

But if you do not know which layers you have, you probably do not have a context system yet. You have scattered artifacts.

That is fixable. It just requires admitting the current setup is usually more improvised than people want to say out loud.

The real advantage is compounding reuse

This is the part that should matter most to operators. Once a context layer exists, you stop re-explaining the company to every workflow. You stop rebuilding the same setup across tools. You stop depending on a senior person's memory to keep outputs aligned. You start getting leverage that compounds because each new workflow inherits a better substrate.

That is the game: not better prompts, not trend-chasing around whichever model demo went viral this week, but a stronger operating environment. If your team is already feeling the pain of prompt glue, shallow outputs, or AI systems that act smart until they hit company-specific reality, the next step is not another prompt library. It is to build the context layer underneath the work.

That is where the durable advantage is going to come from. If you want help turning scattered docs, workflow rules, and business context into an actual operating layer, talk to Deadwater. The companies that win this next phase will not be the ones with the loudest AI strategy. They will be the ones that finally gave the machine a map.

Ready to learn more?

Book a demo and we will walk you through what a Content OS looks like in practice.