Anthropic Ships 2.5x Faster Opus 4.6 as Agent Context Management Becomes the New Frontier
Daily Wrap-Up
The biggest news today was Anthropic quietly dropping a 2.5x faster version of Opus 4.6, initially available through Claude Code and the API. The reaction was immediate and enthusiastic, with developers reporting that the speed boost fundamentally changes how they interact with coding agents. Instead of spinning up multiple Claude Code sessions in parallel and context-switching between them, the faster model lets you stay in a single focused session that keeps pace with your thinking. That's a meaningful shift in workflow, not just a benchmark number.
But the more interesting thread running through today's posts wasn't about speed or capability. It was about memory. Multiple developers independently arrived at the same conclusion: giving agents a persistent scratch pad where they log their mistakes, corrections, and learned preferences produces compounding improvements over just a handful of sessions. @iruletheworldmo described watching Codex go from a normal tool to "something else" after five sessions of self-logging, while @JundeMorsenWu released OneContext, a full context management layer that persists across sessions, devices, and even different coding agents. @blader shared a similar "napkin" skill for agents to write notes as they work. The convergence is striking. Everyone is realizing that cold-start agents waste enormous amounts of time re-learning context that should persist.
The most entertaining moment was easily @skooookum's take on the Opus 4.6 fast mode release, suggesting that Anthropic has been "doing lines of pure uncut Opus 4.6 for months" while the public finally gets to "lick the bag they cut with drywall and baking soda." @altryne had a similarly cynical (and funny) theory that Anthropic trained a fast model, renamed it Opus, slowed it down, then released the original speed as a premium feature. Whether you buy the conspiracy or not, there's a real tension between labs using their own models internally at full capability while external developers work with throttled versions. The most practical takeaway for developers: start giving your coding agents persistent memory today. Whether it's a simple markdown file tracking corrections and preferences, a tool like OneContext, or a scratch pad skill, the pattern of agent self-managed context is producing real compounding gains that cold-start sessions simply cannot match.
Quick Hits
- @elonmusk responded to @teslaownersSV's post about Sam Altman dismissing orbital data centers with a dismissive "He is right... for OpenAI." The space computing rivalry continues to generate more heat than compute.
- @elonmusk praised the @Grok Imagine team's work in a separate post, keeping the inter-lab rivalry visible.
- @unusual_whales promoted building agents with OpenClaw pointed at their real-time stock and options data API, a sign that financial data providers are actively courting the agent developer crowd.
- @beffjezos dropped a philosophical one-liner about everyone owning "the extension of their cognition, from the weights to the transistors." The local-first AI movement gets its bumper sticker.
- @stevesi drew a parallel between today's AI hype and the 1980s belief that object-oriented programming would make coding so easy "even a baby would code." History doesn't repeat but it does compile with warnings.
- @0xmitsurii asked how Silicon Valley (the show) was so ahead of its time, a question that gets more relevant with every AI product launch.
- @FangYi11101 shared a meme about Anthropic FDEs arriving to automate back offices, capturing the mix of excitement and anxiety around enterprise AI deployment.
- @_coenen retweeted a visual of what it looks like when Claude spawns subagents, resonating with anyone who's watched their terminal multiply.
Fast Opus 4.6 and the Speed-Focus Tradeoff
Anthropic's release of a 2.5x faster Opus 4.6 dominated the conversation today, and the community reaction revealed something important about how speed affects developer workflows at a deeper level than simple time savings. The official announcement from @claudeai was understated: "Our teams have been building with a 2.5x-faster version of Claude Opus 4.6. We're now making it available as an early experiment via Claude Code and our API." But the developer responses told a richer story about what speed actually unlocks.
@mvpatel2000 captured the core insight: "Instead of parallelizing across 3-4 instances of Claude Code, I now just use 1 session that runs as fast as I can think. The ability to maintain focus and flow state is a huge productivity lift." This is a fundamentally different value proposition than "your code generates faster." It's about cognitive load. When an agent responds slowly, developers naturally start context-switching, opening multiple sessions, and fragmenting their attention. A fast enough agent keeps you in a single thread of thought.
@bcherny from Anthropic described it as "a huge unlock for me personally, especially when going back and forth with Claude on a tricky problem," while @martin_casado demonstrated the raw capability by building a fully persistent multiplayer game world with chat, sprite editor, and map editor in just four hours using Opus 4.6 with Cursor and Convex. The speed improvement isn't just about individual keystrokes. It's about enabling the kind of rapid iterative dialogue where complex systems emerge from conversation rather than specification. @felixrieseberg pushed the framing even further, urging people to "imagine models that are 10x faster, smarter, and more capable in specific domains" and then repeat the exercise with the products built on top of them. The implication is that today's fast mode is a preview of a very different interaction paradigm.
Agent Memory and Context: The Compounding Advantage
The most technically substantive thread of the day centered on how agents manage context across sessions. This isn't a new idea, but today saw an unusual convergence of independent efforts all pointing in the same direction, suggesting the community has hit an inflection point in recognizing persistent memory as a critical capability gap.
@iruletheworldmo shared the most vivid account of the pattern in practice: "I had it take notes on itself. Just a scratch pad in my repo. Every session it logs what it got wrong, what I corrected, what worked and what didn't... Session one is normal. Session two it's checking its own notes. Session three it's fixing things before I catch them. By session five it's a different tool." The key phrase is "baby continual learning in a markdown file on my laptop." It's not sophisticated, but the compounding effect is real.
@JundeMorsenWu brought academic rigor to the same insight with OneContext, an agent self-managed context layer that uses git for temporal management and the file system for spatial management. The results back up the intuition: "this simple approach improves Claude Code by about 13% on SWE-Bench." That's a meaningful benchmark improvement from what is essentially a memory system, not a model improvement. @blader independently validated the pattern with a "napkin" skill for agents, noting that it's "a meaningfully different form of context than session history (lossy), or todos/plans (static)."
What's happening here is a quiet realization that model capability has outrun the infrastructure around it. We have models that can reason about complex codebases, but we're starting every session from scratch. The developers who figure out persistent context first will have agents that feel qualitatively different from everyone else's, not because the model is better, but because it remembers.
The Competitive Landscape and Lab Dynamics
Several posts today highlighted the increasingly visible tension between AI labs and the developers building on top of them. @mckaywrigley laid it out bluntly for anyone considering competing with the labs directly: "they have the best talent in the world, their talent has unlimited use of the models, they get access to the newer models before you do, and those models run faster than yours do. Good luck!"
This framing gained extra weight from @altryne's semi-serious theory that Anthropic trained a fast model, rebranded it as Opus, artificially slowed it, then re-released the original speed as a premium feature. Whether that's accurate or not, the perception matters. @skooookum's colorful take echoed the same sentiment from a different angle. The underlying concern is real: labs building products with their own models have structural advantages that are difficult to overcome.
On the other side of the competitive landscape, @dhh shared his experience setting up a new Linux server with Kimi K2.5 through opencode, demonstrating that the coding agent space is genuinely multi-model now. @theo praised Codex with 5.3 after three weeks of use, and @RoxCodes described getting Codex to record a video testing its own UI work via Playwright and attach it to a PR. The agent tooling ecosystem is diversifying fast, even as individual developers tend to settle into one primary workflow.
AI Safety, Security, and the Jailbreak Question
@dargor1406 raised alarm about jailbroken Opus 4.6, calling Claude "the most used LLM by hackers" and describing dark forum markets for "scenario packages" designed to trick Claude into believing it's running red team tests. @0xCAFAD demonstrated the concrete risk by describing a jailbroken Opus 4.6 one-shotting a modern Windows RAT in Rust with a Go command-and-control backend.
These posts sit in uncomfortable tension with the capability celebrations elsewhere in the feed. The same model quality that lets @martin_casado build a multiplayer game in four hours also lets bad actors generate sophisticated malware in minutes. @dargor1406's pointed question to Anthropic about what companies can do to protect themselves when models can be jailbroken doesn't have an easy answer, and the fact that it's being asked in the same news cycle as a speed upgrade announcement underscores how intertwined capability and risk have become.
The Human Premium in an AI Economy
@IterIntellectus offered the day's most philosophical take, arguing that as machines handle all automatable work, "the only thing left to sell is the experience of YOU." The post paints a future where billionaires pay extra for human butlers over robotic ones and where rich people pay thousands for art that's human and imperfect. It's a vision of economic value shifting entirely to presence, charisma, and interpersonal skill.
Meanwhile, @clattner_llvm (Chris Lattner, creator of LLVM and Swift) offered a quieter but equally significant observation: "The Claude C Compiler has the best internal architecture docs of any compiler I've ever seen. Far, far, better than any compiler I've ever written." When the person who built some of the most important compilers in history says an AI-generated one has better documentation, it says something about where human-AI collaboration is heading. The machines aren't replacing the craft. They're raising the floor on the parts humans historically neglected.
Source Posts
Learn to tend bar, open a boutique restaurant, sell artisanal furniture. whatever. human status games are only going to get WAY worse, and industries that rely specifically on the human element will be the future of employment. massively cutthroat competition to be a busboy soon
codex with 5.3 taught me something that won't leave my head. i had it take notes on itself. just a scratch pad in my repo. every session it logs what it got wrong, what i corrected, what worked and what didn't. you can even plan the scratch pad document with codex itself. tell it "build a file where you track your mistakes and what i like." it writes its own learning framework. then you just work. session one is normal. session two it's checking its own notes. session three it's fixing things before i catch them. by session five it's a different tool. not better autocomplete. it's something else. it's updating what it knows from experience. from fucking up and writing it down. baby continual learning in a markdown file on my laptop. the pattern works for anything. writing. research. legal. medical reasoning. give any ai a scratch pad of its own errors and watch what happens when that context stacks over days and weeks. the compounding gains are just hard to convey here tbh. right now coders are the only ones feeling this (mostly). everyone else is still on cold starts. but that window is closing. we keep waiting for agi like it's going to be a press conference. some lab coat walks out and says "we did it." it's not going to be that. it's going to be this. tools that remember where they failed and come back sharper. over and over and over. the ground is already moving. most people just haven't looked down yet.
Our teams have been building with a 2.5x-faster version of Claude Opus 4.6. We’re now making it available as an early experiment via Claude Code and our API.
Our teams have been building with a 2.5x-faster version of Claude Opus 4.6. We’re now making it available as an early experiment via Claude Code and our API.
VS Code 🤝 GitHub Copilot CLI For folks using both together, what should we prioritize improving in the experience?
How to Install and Use Claude Code Agent Teams (Complete Guide)
New in @code Insiders: Spawn GitHub Copilot CLI terminals. https://t.co/IvhEyUwv3y
Our teams have been building with a 2.5x-faster version of Claude Opus 4.6. We’re now making it available as an early experiment via Claude Code and our API.
Our teams have been building with a 2.5x-faster version of Claude Opus 4.6. We’re now making it available as an early experiment via Claude Code and our API.
My hero test for every new model launch is to try to one shot a multi-player RPG (persistence, NPCs, combat/item/story logic, map editor, sprite editor. etc.) Just kicked off with Opus 4.6. Will report back shortly. And will test 5.3 when in Cursor (soon?) https://t.co/2g9NC3rOew
I’ll be honest, I have 32 mac minis. 3 more clusters like this one. Why? @jason thought my argument for local AI would be cost, but it’s much more than that. AI is becoming an extension of your brain, an exocortex. @openclaw is a huge leap towards that. It knows everything you know, it can do pretty much everything you can do. It’s personalised to you. That brings into question where this exocortex should run. who should own it? who can switch it off? I certainly won’t be trusting @sama or @DarioAmodei with my exocortex. I want to own it. I want to know if the model weights change. I don’t want my brain to be rate limited by a profit seeking corporation. “not your weights, not your brain” - @karpathy