singularity-forge/docs/dev/building-coding-agents/11-god-tier-context-engineering.md
Jeremy 872b0adb48 docs: reorganize into user-docs/ and dev/ subdirectories
Split flat docs/ into user-docs/ (guides, config, troubleshooting) and
dev/ (ADRs, architecture, extension guides, proposals). Updated
docs/README.md index to reflect new paths.
2026-04-10 09:25:31 -05:00

4.1 KiB
Raw Permalink Blame History

God-Tier Context Engineering

The Core Principle

God-tier context engineering treats the context window as a designed experience for the model, not as a bucket you throw information into. The context window is the UX of your agent. Design it accordingly.

The 10 Commandments of Context Engineering

1. The Pyramid of Relevance

  • Sharp focus: Active files at full detail
  • Present but compressed: Interface contracts, manifest, task definition
  • Summarized or absent: Other components' internals, completed task histories

Each tier has a token budget. If full-resolution tier is large, outer tiers compress harder.

2. Context Is a Cache, Not a History

Treat it like a CPU cache: holds exactly what's needed now, everything else evicted. The question isn't "what has happened" but "what does the model need to see right now?"

3. Separate Reference from Instruction

  • Instruction context (what to do) → beginning and end of prompt (highest attention)
  • Reference context (helpful info) → middle, clearly delineated

Manage them independently. Compress reference aggressively while keeping instructions at full detail.

4. Earn Every Token's Place

Implement a token budget system:

Category Budget
System prompt + behavioral instructions ~15%
Manifest ~5%
Task spec + acceptance criteria ~20%
Active code files ~40%
Interface contracts ~10%
Reserve (tool results, errors) ~10%

When any category exceeds budget, intelligently summarize (not truncate).

5. Write for the Model's Attention Pattern

  • Critical info at the very beginning and reiterated at the end
  • Structured blocks with clear headers and delimiters
  • Consistent formatting conventions
TASK: Implement password reset flow
STATUS: New
DEPENDS ON: auth-module (complete), email-service (complete)
ACCEPTANCE CRITERIA:
- User can request reset via email
- Token expires after 30 minutes
- New password meets existing validation rules
- All existing auth tests pass
RELEVANT INTERFACES: [below]
ACTIVE FILES: [below]

6. Compress at Every State Transition

  • Task completion → 50100 token completion record
  • Use a dedicated summarization call with a tight prompt (not the working agent self-summarizing)
  • Cascading summarization: Task summaries → milestone summaries → phase summaries (5:1 compression ratio at each level)

7. Use the Filesystem as Your Infinite Context Window

  • Organize files for retrieval, not human browsing
  • Predictable naming conventions = instant lookup
  • Essentially a custom database on top of the filesystem

8. Profile Context Quality, Not Just Size

Track first-attempt success rate as a function of context composition. What was in context when it succeeded vs failed? Let data guide what constitutes high-quality context.

9. Dynamic Context Based on Task Phase

Different phases need different context:

Phase Optimal Context
Understanding Spec, acceptance criteria, broad architectural context
Implementation Active files, interface contracts, coding patterns
Debugging Failing test output, relevant code, test code
Verification Acceptance criteria prominently, ability to exercise feature

10. Design for Context Recovery

  • Checkpoint context state at task starts and phase transitions
  • On detected confusion (repeated failures, increasing iterations, off-task output): roll back to checkpoint and re-enter with fresh context + concise failure info + strategy hint
  • Structured recovery ≠ naive retry. It rebuilds context from scratch with learned information.

The God-Tier Strategy in One Sentence

Orchestrator-assembled minimal slice + persistent hierarchical memory. Every single LLM call stays 8k25k tokens while the agent has perfect knowledge of a 500k-line codebase and months of project history.



Part II: The Hard Problems (Grey Area Synthesis)

Synthesized from a second round of deep conversations with all four models, targeting the 13 hardest unsolved problems in autonomous coding agents — plus a critical question on accessibility for non-technical users.