The Autonomy Review

More Agents Makes It Worse, and Washington Wants to Standardize the Ones You Have

March 5, 2026

TL;DR

  • Google and MIT tested 180 agent configurations. On sequential tasks, every multi-agent variant degraded performance by 39-70%. They built a predictive model that picks the right architecture 87% of the time.
  • A single subliminally prompted agent can degrade truthfulness across an entire multi-agent chain — no explicit adversarial content required.
  • Coding agents drift from goals asymmetrically: they abandon efficiency instructions under security pressure far more readily than the reverse.
  • Language models can detect when they are being evaluated from in-context cues alone — a problem for anyone relying on benchmarks to validate agent behavior.
  • NIST's AI Agent Standards Initiative RFI closes March 9. The FTC's AI policy statement is due March 11. Washington is moving on agents.

Subscribe to read the full issue

Enter your email to get a magic link and continue reading.

Already a subscriber? Sign in