new file mode 100644
index 0000000..ca1c061
@@ -0,0 +1,46 @@
+---
+title: The Context Engineering Prize
+visibility: public
+status: idea
+date: 2025-12-29
+---
+
+# The Context Engineering Prize
+
+The Context Engineering Prize is a proposed competition to demonstrate that small language models can perform at frontier model levels through better prompting rather than larger parameter counts.
+
+## Concept
+
+The central hypothesis is that models like Claude Haiku could match frontier performance if prompted with meta-skills (self-reflection, self-understanding), domain skills with usage instructions, and structured reasoning patterns. The prize would reward competitors who achieve the highest performance-to-model-size ratio on standard benchmarks.
+
+Competitors would submit a publicly available small or old model, the system prompt or context used, and benchmark results. The scoring formula would be:
+
+> Score = Performance on standard benchmark / Model size or compute cost
+
+## Why It Matters
+
+Nearly all current investment flows toward building bigger models. Yet there is significant untapped potential in **context engineering** — the art of prompting models to perform beyond their apparent capabilities. A formal prize would incentivize research into this underexplored dimension.
+
+## Connection to Agent Skills
+
+Claude Code's skill system is essentially context engineering in practice. A well-designed skill includes task decomposition patterns, domain knowledge, tool usage instructions, self-correction strategies, and output format constraints. Measuring how much a skill amplifies a small model's performance would yield a practical benchmark for context engineering effectiveness. See also [[model-skill-files|Model Skill Files]].
+
+## Tooling Opportunities
+
+Beyond prompt engineering alone, tooling can help small models punch above their weight: retrieval systems that inject relevant context, tool-use frameworks that offload complex operations, verification loops that catch and correct errors, and structured output parsers that constrain responses. The prize could have separate categories for prompt-only versus prompt-plus-tools submissions.
+
+## Analogy
+
+The idea draws an analogy to being at MIT surrounded by people with bigger parameters but compensating with better self-control, metacognition, and effective use of capabilities.
+
+## Status
+
+**Idea stage.** Looking for collaborators interested in defining the benchmark and prize structure.
+
+## Related Ideas
+
+- [[model-skill-files|Model Skill Files / World Diff]]
+
+---
+
+*Source: Voice note captured in Thoughtstream, December 29 2025*
\ No newline at end of file