Harness around an LLM delivers multi-fold gains: what changed after a year with Claude Code
After a year and a half with Claude Code, an engineer found that the main quality lever for an LLM is not new model versions but the harness around it: system p

◐ Listen to article
After a year and a half with Claude Code, an engineer found that the main quality lever for an LLM is not new model versions but the harness around it: system prompt, tools, context, skills, hooks, permissions and memory. Each layer of that harness delivers multi-fold gains, while switching models brings a noticeable but limited result.