ASJSR

American Scholarly Journal for Scientific Research

The Art of Omission: How Context Engineers Are Rewriting the Rules of AI Memory

By Sarah Mitchell ·
The Art of Omission: How Context Engineers Are Rewriting the Rules of AI Memory

Context is the new code.

Three years ago, developers tweaked hyperparameters to coax better results from neural networks. One year ago, they wrote elaborate system prompts. Today, the most effective practitioners are doing something counterintuitive: they are building systems that strategically forget. This is context engineering — and the field is maturing faster than most teams realize.

The future of AI performance belongs to those who master what they leave out.

What Prompt Engineering Got Wrong

Prompts are a single message.

Prompt engineering promised control through phrasing. If you wrote the question correctly, the model responded correctly. Thousands of courses, GitHub repos, and Twitter threads built entire careers on this premise — and the premise was incomplete.

Context is an entire ecosystem.

The model doesn't just respond to your latest instruction. It responds to everything it can see: system messages, conversation history, retrieved documents, tool outputs, previous agent steps, memory summaries, and the quiet pressure of whatever came before. The token window is not a prompt. It's an environment — and environment shapes behavior.

Why "More Context" Fails

The instinct when an AI agent makes mistakes is to add information. Clarify the instructions. Add more examples. Include more retrieved chunks.

This is almost always wrong.

  • Context windows degrade under load — attention dilutes across thousands of tokens
  • Long contexts force models to split reasoning across vast distances of text
  • Facts buried in the middle of long contexts are systematically underutilized
  • More tokens mean more noise, not more signal

Researchers call this the "lost in the middle" problem: information sandwiched between a long preamble and a long suffix gets ignored by the model at rates that would shock most developers. More context doesn't mean better reasoning. It means harder reasoning about worse information.

The Three Disciplines of Context Engineering

Context engineering lives at the intersection of three distinct disciplines, none of which are about writing better prompts:

  • Information architecture — deciding what facts belong in the context window, in what order, and with what framing to guide attention
  • Memory systems — designing how an agent accumulates, summarizes, and retrieves knowledge across conversations and sessions without saturating working context
  • Attention shaping — structuring content so the model's attention concentrates on the right tokens at the right moment in the reasoning chain

Each discipline requires its own design patterns, tooling, and evaluation criteria. Teams that treat them as one problem — "just write a better prompt" — consistently underperform teams that address them separately.

The Omission Principle

The most underrated skill in context engineering is deletion.

Every token you include costs the model attention. Every piece of irrelevant information crowds out something relevant. The best context engineers treat the token window like a financial budget: every item must earn its place, and anything that doesn't carry its weight gets cut.

This feels wrong at first. The fear is that removing information will cause the model to miss something critical. In practice, the opposite happens:

  • Focused, tight contexts produce sharper, more reliable reasoning
  • Bloated, comprehensive contexts produce hedging, confusion, and hallucination
  • Models aren't smarter when they have more information — they're smarter when they have the right information

The omission principle doesn't mean stripping out necessary context. It means being ruthless about what counts as necessary — and the bar is higher than most teams set it.

From Prompts to Pipelines

The shift from prompt engineering to context engineering is a shift from documents to systems.

A prompt is static: written once, applied uniformly. A context pipeline is dynamic: responsive to the current state of the conversation, the task, and the agent's accumulated memory. Context engineers don't just write instructions. They build machinery that decides what instructions to surface, when to surface them, and what to leave in storage.

  • Static retrieval (RAG) gives way to dynamic retrieval, where what gets pulled depends on current task state rather than query similarity alone
  • Fixed system prompts give way to assembled context, where instructions compose from modular pieces selected for current agent needs
  • Flat memory gives way to hierarchical memory, with working context, episodic summaries, and long-term knowledge at different levels of compression
  • Monolithic agents give way to context-aware agent graphs, where each node in the pipeline receives only what it needs to function

This is fundamentally a software engineering problem, not a writing problem.

The Agentic Amplification Effect

Autonomous agents amplify the stakes of context engineering dramatically.

A single bad prompt in a chatbot produces one bad response. A single bad context strategy in an agent produces cascading failures across dozens of tool calls, memory writes, and decision branches. The compounding nature of agentic execution means that context errors don't stay local — they propagate.

The best agentic systems treat context as a first-class resource with explicit allocation policies:

  • What does this agent need to know right now?
  • What can it look up on demand without loading into working context?
  • What must persist across the full task lifecycle?
  • What can it safely discard after each step?

These questions aren't answered in a prompt. They're answered in architecture — before the agent runs, not after it fails.

Measuring What Matters

Teams making this transition report a consistent pattern in how they measure success:

  • They spend less time writing long system prompts
  • They spend more time designing what enters the context window and in what order
  • They treat context construction as a versioned, testable, observable engineering concern
  • They shift from measuring output accuracy to measuring reasoning quality

The measurement shift matters most. Accuracy metrics tell you if the answer is right. Reasoning quality metrics tell you why — which means you can diagnose and fix failures instead of just retrying them.

The Practical Starting Point

For teams beginning this transition, three changes produce immediate results:

  • Audit your current context — log the full context your model receives and identify what percentage of tokens are actually necessary for the task at hand
  • Separate retrieval from reasoning — pull documents and tool outputs in one step, then filter and compress before loading into the reasoning context
  • Build context tests — treat context construction as testable logic, not as a prompt you iterate on manually until it feels right

None of these changes require new models or new infrastructure. They require a different mental model of what the problem actually is.

The Real Competitive Edge

Every company building with AI is writing prompts.

Few companies are engineering context. Fewer still are doing it systematically, with proper tooling, evaluation, and iteration loops. The gap between these two approaches will define which teams ship reliable, scalable AI products — and which teams spend their engineering cycles firefighting hallucinations and inconsistent behavior that no amount of prompt rewriting seems to fix.

Context is the new code. Master what you leave out.

S

Sarah Mitchell

Sarah Mitchell is a software engineer and AI systems researcher specializing in production agentic AI architectures and context-aware pipeline design.