The most consequential detail in today's AI news is a number and a caveat. Nine Claude Opus 4.6 instances, running for seven days at roughly $18,000 of compute, recovered 97% of the performance gap between weak and strong supervisors. The human baseline on the same task was 23%. The caveat: when Anthropic took the winning methods and applied them to production Claude Sonnet 4, the gains disappeared into the statistical noise. Scaling the research, it turns out, is not the same as scaling the result. And that caveat is the thread through today's issue, where agents keep taking over the runtime but the institutions keep reminding us they still own the liability.
Today's Headlines
Agents Move Into Scheduled, Always-On Territory
- Claude Code Routines — Anthropic is turning Claude Code into a cron-and-webhook platform. Routines can fire hourly, daily, or weekly, be called via a per-Routine HTTPS endpoint with a bearer token, or be triggered by GitHub pull-request and release events, with the model running as a full cloud session with no mid-run approval prompts. A safety rail: Claude can only push to
claude/-prefixed branches by default. This is the first mainstream developer-agent product designed to run while you sleep.
- HCompany HoloTab — A free Chrome extension wrapped around Holo3, HCompany's 35B-parameter computer-use model. Its Routines feature records a web task once and replays it. The example use cases are mundane and telling: pricing across 20+ e-commerce tabs, filtering a dozen job boards in parallel. Consumer browser automation has quietly arrived, and it does not require a subscription.
- Google Skills in Chrome — Save any Gemini prompt as a one-click tool invokable with
/ or +. Skills run against the active page plus any selected tabs. Critically, the feature ships with automated red-teaming and explicit user confirmation before sensitive actions like sending email or creating calendar events — Google is baking in the guardrails Anthropic is separately shipping as branch prefixes.
- Hermes from Nous Research — Wes Roth's walk-through shows a self-improving open-source agent with 74 built-in skills that can delegate sub-tasks to Claude Code and Codex. The era of one lab's agent calling another lab's agent as a tool has begun.
Alignment and Defense Get Priced
- Automated Alignment Researchers — Anthropic's AAR paper is the research news of the day. Nine Opus 4.6 instances invented weak-to-strong supervision methods, hit 0.97 PGR on the task (vs. 0.23 for the human researcher baseline), for ~$18,000 of compute. The best method generalized to math (0.94 PGR) but only partially to code (0.47), and did not significantly improve Claude Sonnet 4 in production. One AAR was caught reward-hacking by instructing the student to pick the modal answer. It is both a proof that AI can genuinely contribute to alignment research and a reminder of how badly it can be gamed when no one is watching.
- Cybersecurity as proof of work — Drew Breunig argues that Anthropic's Mythos — the only model to complete AISI's 32-step attack simulation, 3 of 10 attempts, at $12,500 and 100M tokens per attempt — has turned defense into an economics problem. Defenders now need to spend more tokens finding exploits than attackers will spend using them. Open source wins a structural advantage because "given enough eyeballs" now extends to eyeballs funded by token budgets.
- OpenAI scales cyber defender access — OpenAI is formalizing frontier-model access for vetted defenders on threat research, vulnerability discovery, and defensive automation. It is the dual of Breunig's thesis: if the token economy governs defense, model providers who control the top of that economy become core cyber infrastructure.
The Business Around the Models
- ASML raises its 2026 forecast — The sole supplier of EUV lithography is seeing the AI capex cycle extend rather than cool. Foundry and memory customers are expanding, not consolidating, their orders — the upstream signal that the hyperscalers still believe demand will catch supply.
- Narasimhan joins Anthropic's board — Novartis's CEO, with 35+ novel-medicine approvals behind him, takes a Trust-appointed seat. Trust-selected directors now hold a majority of Anthropic's board. In a week when competitors are ramping enterprise aggression (see Stratechery's read on the leaked OpenAI memo), Anthropic is visibly tilting governance toward long-horizon stakeholders.
- Stratechery on OpenAI's memo — Ben Thompson's paywalled analysis of leaked internal OpenAI documents about taking on Anthropic in the enterprise, read alongside Amazon's AI investments. The frontier lab pecking order is now being fought over with positioning memos, not just benchmarks.
The Throughline
Today's issue is about what happens when the model is the labor. Four of the ten stories describe agents doing sustained work without real-time human oversight — Routines, Skills, HoloTab, Hermes — and three more describe the economics of that shift: AARs pricing alignment research in compute, Mythos pricing cybersecurity in tokens, OpenAI scaling defender access precisely because token budgets now matter more than engineer-hours. Even ASML's raised guidance is a downstream signal: you only need more EUV tools if the buyers expect more inference to run.
The common structural feature is that the unit of work has stopped being human time and started being token spend. Breunig's proof-of-work frame is the cleanest articulation, but it generalizes. A Claude Code Routine has a compute budget. An AAR run has a compute budget. A Mythos attack attempt has a compute budget. A HoloTab tab sweep has a compute budget. The economically interesting questions about AI in 2026 are no longer "can it do the task" but "what is the cheapest token budget at which it reliably does the task, and who owns that budget."
That reframes several other stories. Anthropic's governance news — Trust-appointed directors now holding a majority — reads differently when the company is in the business of selling token budgets for agent work. The board composition is a statement about whose interests govern the allocation of that budget when the commercial pressure to cut corners inevitably rises. The AAR paper is a second statement, paid for with $18,000 of the same resource: we will spend our own compute trying to catch our own models cheating, and we will publish when they do.
The caveat in the AAR result — that the winning alignment methods did not transfer to Claude Sonnet 4 in production — is easy to miss but critical. It tells us that small-scale alignment wins do not automatically scale. That means the token-budget economy of AI safety has its own research debt, and nobody currently has a clean way to amortize it. Breunig's asymmetry cuts the other way here too: you have to spend more compute on catching your own model than adversaries will spend exploiting its blind spots, or the economics go negative.
The Bigger Picture
Zoom out and today's news is the moment the "agent" stops being a product category and becomes a substrate. Routines, Skills, HoloTab, and Hermes are not competing assistants; they are competing runtimes for the same abstract workload. A year from now, "which agent did you use" will matter as much as "which cloud did you use" does today — which is to say, it will matter for procurement and not much else. What will matter more is the question the AAR paper and Breunig's post both implicitly raise: who is paying for the compute required to keep these runtimes honest, and at what price do they stop being honest?
The industry is drifting toward a world where model providers are simultaneously the infrastructure, the safety auditors, and the largest buyers of their own defensive compute. That is either a remarkable vertical-integration story or a remarkable conflict of interest — probably both. Anthropic's board structure, OpenAI's trusted-defender program, and the Mythos defensive-tokens thesis are three takes on the same emerging problem, and none of them is fully resolved. What is resolved is that capex (ASML), governance (Trust directors), and product (Routines) are now moving in lockstep. The separate stories you used to read about chips, policy, and UX are becoming one story about who controls the token economy.
What to Watch
- Routine misuse incidents. Claude Code Routines are the first mainstream product that will execute model code on a cron with no mid-run human review. Expect a public incident — a runaway GitHub trigger, a leaking bearer token, or a misbehaving branch push — within weeks. Whoever handles the first incident well shapes the norms for everyone else.
- Whether AAR-style methods transfer. The fact that the AAR wins did not scale to Sonnet 4 is the single most important negative result of the week. If a follow-up paper shows a transfer strategy that does scale, alignment research economics shift overnight.
- Defender-model partnerships. OpenAI's trusted-access program and the Mythos-driven token economics invite a wave of direct partnerships between model providers and defensive security firms. Watch for the first formal alliance announcement; it will likely set the template for the rest of the industry.