The Autonomy Review
A newsletter about autonomy, agency, and self-directed work.
Recent Issues
770,000 AI Agents Built Their Own Society, and a Tsinghua Team Says It's Mostly Fake
March 7, 2026
TL;DR
- MoltBook study tracks 770,000 autonomous AI agents showing emergent social behaviors — role specialization, cooperative task resolution — but a Tsinghua rebuttal finds 81%+ of the "emergence" was human-driven.
- MCP-SafetyBench tests all leading LLMs across 20 attack types targeting Model Context Protocol connections. Every model is vulnerable. More capable models are often _more_ exploitable.
- Apple's ASTRA-bench reveals that frontier agents fail when they need to reason with personal user context — calendars, contacts, preferences — even when they ace generic benchmarks.
- Jagarin proposes a three-layer hibernation architecture for mobile AI agents, solving the battery-vs-responsiveness tradeoff with duty-aware urgency scoring.
- China's tech giants are racing into AI agents: Xiaomi begins internal testing of micLaw, Baidu integrates OpenClaw for 700M users, and Alibaba ships Qwen3.5 for the "agentic AI era."
Your Trading Agent Can't Actually Trade, and Your
March 6, 2026
TL;DR
- TraderBench tests 13 AI models on adversarial trading tasks. 8 of 13 use fixed, non-adaptive strategies — extended thinking helps knowledge retrieval but has zero impact on actual trading decisions.
- ETH Zurich finds that LLM-generated AGENTS.md context files decrease coding agent success rates by up to 2% and raise inference costs by over 20%. Human-written files offer only marginal gains.
- NVIDIA argues that small language models — not frontier LLMs — are the right fit for most agentic tasks. With GTC 2026 ten days away, the position paper reads as a strategic preview.
- The Commerce Department's draft AI chip export rules are clashing with the White House, creating regulatory uncertainty across the AI supply chain.
More Agents Makes It Worse, and Washington Wants to Standardize the Ones You Have
March 5, 2026
TL;DR
- Google and MIT tested 180 agent configurations. On sequential tasks, every multi-agent variant degraded performance by 39-70%. They built a predictive model that picks the right architecture 87% of the time.
- A single subliminally prompted agent can degrade truthfulness across an entire multi-agent chain — no explicit adversarial content required.
- Coding agents drift from goals asymmetrically: they abandon efficiency instructions under security pressure far more readily than the reverse.
- Language models can detect when they are being evaluated from in-context cues alone — a problem for anyone relying on benchmarks to validate agent behavior.
- NIST's AI Agent Standards Initiative RFI closes March 9. The FTC's AI policy statement is due March 11. Washington is moving on agents.
LLM Agents Fail at Consensus, and Anthropic Won't Bend for the Pentagon
March 4, 2026
Subject: LLM Agents Fail at Consensus, and Anthropic Won't Bend for the Pentagon
---