This week’s picks are about making coding agents operational: pick the right harness boundaries, then raise verification frequency so high-velocity diffs don’t turn into variance. The recurring move is to treat context and guardrails as first-class artifacts you can validate, version, and iterate on.

  • Emerging Patterns for Coding with Generative AI — DevCon Fall 2025 (Lada Kesseler) — The durable shift is from “better prompting” to context management: capture decisions into reloadable knowledge docs, keep instructions tight to avoid context rot, and use specialist agents when focus matters. Two tactics worth stealing immediately: Semantic Zoom (zoom out, then drill in) and the “feedback flip” where you force a reviewer pass before you trust a diff — a concrete way to operationalize a checker.

  • How to Make the Best of AI Programming Assistants (Dave Farley) — A clean frame: AI increases the frequency of changes, so teams start under-sampling their own output. The fix isn’t more careful reading — it’s faster, more frequent signals: keep diffs small, run CI on every meaningful step, and treat tests/checks as the source of truth. This aligns with Thin Slices and Tight Loops: shrinking batch size is what makes higher cadence safe.

  • Pi Coding Agent — Open-Source Harness Customization — The argument for Pi isn’t “smarter prompts,” it’s harness leverage: hooks, UI widgets, and a customizable event loop let you enforce task lists, block dangerous commands, and make behavior observable. The trade-off is explicit: the default posture is closer to “YOLO mode,” so reliability comes from what you build on top — a useful complement to the more polished defaults in Claude Code and the general framing in Agentic Harness.

  • OpenClaw vs Claude Code — Reach vs Control — A practical decision axis: always-on reach (chat-app access, proactive assistance) tends to imply broader permissions and a bigger blast radius; repo-scoped deep work trades accessibility for safer boundaries and lower operational overhead. The actionable pattern is role-splitting: isolate the high-reach agent on a sacrificial host, and keep the high-impact coding loop bounded — a harness choice that maps cleanly onto Orchestrator vs. Executor and Agentic Harness.

  • Context as Code — Stop Prompting, Start Engineering (Dru Knox) — Treat context like software: validate formats, run scenario-based evals with rubrics, add repo-level “integration evals” to catch the “too much context” dumb zone, and mine logs to drive improvements. The most useful takeaway is the lifecycle mindset: context needs CI, observability, and an update mechanism or it rots. This pairs naturally with Context Hygiene and Context is a Per-Feature Budget.