(no title)
youraimarketer | 2 months ago
The solutions to these problems are scattered across research papers, framework docs, and production war stories. I collected and synthesized them into a set of "Agent Skills" - structured instructions that agents can load on demand when working on relevant tasks.
7 skills covering context engineering fundamentals:
- \context-fundamentals\: What context actually is (system prompts, tool definitions, retrieved docs, message history, tool outputs) and why context quality matters more than context length
- \context-degradation\: The failure modes - lost-in-middle (10-40% accuracy drop for middle content), context poisoning (hallucinations that compound), context distraction (irrelevant info consuming attention budget)
- \multi-agent-patterns\: Supervisor vs swarm vs hierarchical architectures, when to use each, and the "telephone game" problem where supervisors paraphrase sub-agent responses incorrectly
- \memory-systems\: Why vector stores lose relationship information, when to use knowledge graphs, and how temporal validity prevents outdated facts from conflicting with new ones
- \tool-design\: The consolidation principle (if a human can't say which tool to use, an agent can't either), error messages that enable recovery, response format options for token efficiency
- \context-optimization\: Compaction triggers, observation masking (tool outputs can be 80%+ of token usage), KV-cache optimization
- \evaluation\: Multi-dimensional rubrics instead of single metrics, LLM-as-judge for scale, human review for edge cases
It uses Anthropic's open Agent Skills format. Each skill is a folder with a SKILL.md file containing instructions. Progressive disclosure - agents load only skill names/descriptions at startup, full content loads when activated for relevant tasks.
Works with Claude Code, Cursor, or any agent that supports skills/custom instructions.
Would appreciate feedback, especially from anyone running multi-agent systems in production. What patterns are you seeing that aren't captured here?
No comments yet.