Discussion about this post

User's avatar
MetaCortex Dynamics's avatar

"Fix-precision was 33.7%. Regression-precision was 11.8%." That asymmetry is the diagnostic.

The agent is good at predicting what it will fix because forward justification is pattern matching against the error it just observed. The agent is bad at predicting what it will break because defensive prediction requires modeling the full consequence space of the edit. The asymmetry is not a tuning problem. It is architectural: the agent does not have a structural gate that tests the edit against invariants before promotion.

We built the gate. Ninety governed experiments on a nanochat benchmark. Every proposed edit was evaluated by a structural gate before promotion. The gate does not predict what the edit will break. The gate tests the edit against the measured invariant. If the edit degrades the invariant, the gate rejects. If it does not, the gate accepts. Zero false keeps across 90 experiments.

The difference: their system predicts regressions at 11.8% precision and rolls back on failure. Our system prevents regressions architecturally and never needs to roll back. Prediction is unnecessary when verification is structural.

"The model is rented. The harness is owned." Exactly right. The harness is the durable IP. The next step is making the harness governed: PROPOSE the edit, DECIDE against the invariant, PROMOTE to staging, EXECUTE only after promotion. Four separate phases. No rollback needed because no regression passes the gate.

No posts

Ready for more?