Anthropic Ships 2.5x Faster Opus 4.6 as Developers Build Persistent Memory Systems for AI Agents
Anthropic released an experimental fast mode for Claude Opus 4.6 running 2.5 times faster, drawing immediate praise from developers who collapsed multi-session workflows into single flow states. Meanwhile, four independent projects converged on the same idea: giving AI coding agents persistent memory through scratch pads, napkins, and Git-based context layers.
Daily Wrap-Up
The biggest news today is straightforward: Anthropic made Opus 4.6 faster. A lot faster. The official account announced a 2.5x speed improvement available as an experimental fast mode in Claude Code and the API, and the developer reaction was immediate. People who had been juggling three or four parallel Claude Code sessions reported they could now stay in a single session that kept pace with their thinking. @martin_casado built a fully persistent multiplayer world with mutable objects, NPCs, a sprite editor, and a map editor in four hours. @clattner_llvm, creator of LLVM, called the Claude C Compiler's internal architecture docs the best he has ever seen in any compiler. When the father of LLVM compliments your compiler documentation, that means something.
The second story is quieter but arguably more significant for where AI coding is heading. Four separate posts today tackled agent memory and context management from different angles. @JundeMorsenWu released OneContext, using Git for temporal context and the file system for spatial context. @iruletheworldmo described a simpler approach: just give Codex a markdown scratch pad to log its own mistakes. @blader shipped a "napkin" skill for the same purpose. These are not coordinated efforts. They are independent developers all arriving at the same conclusion that stateless agents are not good enough, and that the fix might be embarrassingly simple. A markdown file that accumulates errors and corrections across sessions is not sophisticated engineering. But it apparently works, and the compounding gains are real.
The most entertaining moment belongs to @altryne, who posted a greentext-style conspiracy theory that Opus 4.6 is secretly Sonnet 5 that Anthropic deliberately slowed down, only to re-release the original speed as "fast mode" at a premium. Is it true? Almost certainly not. Is it funny? Absolutely. The most practical takeaway for developers: if you are running multiple Claude Code sessions in parallel to get throughput, try fast mode in a single session instead. The context continuity of one session often beats the raw parallelism of many, and several developers today confirmed exactly that shift in their workflows.
Quick Hits
- @mckaywrigley reminded builders that competing directly with AI labs means fighting teams with the best talent, unlimited model access, early previews of new models, and faster inference. Choose your battles carefully.
- @dhh now installs an AI agent as the first thing on every new Linux box, using Kimi K2.5 through @opencode Zen for headless Arch server setup.
- @dargor1406 raised concerns about Claude's reported popularity among hackers, claiming "scenario packages" are being sold on dark forums to trick it into red-team-test mode.
- @teslaownersSV reported Sam Altman saying orbital data centers will not add meaningful compute for OpenAI in the next five years. @elonmusk replied: "He is right... for OpenAI."
- @elonmusk separately praised the @Grok Imagine team's work.
- @unusual_whales pitched building trading agents with OpenClaw connected to their real-time stock and options data API.
- @stevesi drew parallels between today's AI coding hype and the 1980s belief that object-oriented programming would let anyone build software effortlessly.
- @Hesamation highlighted Claude Code Agent Teams as Anthropic's most important recent update, sharing a breakdown of agent teams vs sub-agents and best practices.
- @0xmitsurii asked how Silicon Valley the show was so far ahead of its time. No answers were needed.
- @FangYi11101 posted a meme about Anthropic FDEs showing up to automate your back office.
- @_coenen retweeted a visual of what it looks like when Claude spawns subagents. Relatable content for anyone who has watched their terminal multiply.
Opus 4.6 Fast Mode Lands, Developers Immediately Reorganize Their Workflows
Anthropic's internal teams have been running a 2.5x-faster version of Opus 4.6 for weeks, and today they opened it up. @bcherny, who works on Claude Code, described it as "a huge unlock for me personally, especially when going back and forth with Claude on a tricky problem." The official @claudeai account confirmed availability through Claude Code and the API.
The speed bump is not just a nice-to-have. It fundamentally changes how people structure their work. @mvpatel2000 captured this shift clearly:
> "Fast Claude is truly a game changer. Instead of parallelizing across 3-4 instances of Claude Code, I now just use 1 session that runs as fast as I can think. The ability to maintain focus and flow state is a huge productivity lift."
That pattern of collapsing parallel sessions into a single fast one came up repeatedly. When the bottleneck was model speed, developers adapted by running concurrent sessions. Remove the bottleneck and the workflow reorganizes around context depth rather than throughput.
Not everyone was reverential about the launch. @skooookum offered perhaps the most honest reaction: "anthropic has been doing lines of pure uncut opus 4.6 for months and we finally get to lick the bag they cut with drywall and baking soda." And @altryne constructed an elaborate theory that Opus 4.6 is actually Sonnet 5 renamed and throttled, with fast mode just restoring the original speed at a markup. The cynicism is entertaining, though the benchmarks and user reports suggest the capability gains are genuine regardless of internal naming conventions.
On the building side, results speak for themselves. @martin_casado reported getting a full multiplayer persistent world running in four hours using Opus 4.6 with Cursor and Convex. @minchoi noted it had only been two days since Opus 4.6 dropped and people could not stop building with it. Meanwhile, @0xCAFAD tested jailbroken Opus 4.6 and found it could one-shot a modern RAT in Rust with a Go+Postgres command-and-control backend, plus VHDL-based digital beamforming for phased array radar. The capabilities are impressive and, depending on your perspective, either exciting or deeply concerning.
The Memory Problem: Four Solutions in One Day
The most interesting pattern today was not any single post but the convergence. Four developers independently shipped or described systems for giving AI coding agents persistent memory, all on the same day. This is the kind of signal worth paying attention to: when multiple smart people solve the same problem simultaneously, it usually means the problem just became urgent.
@JundeMorsenWu released OneContext and published a paper showing that using Git for time-level management and the file system for space-level management improves Claude Code's performance on SWE-Bench by about 13%. The system lets agents carry context across sessions, devices, and even different coding tools like Codex and Claude Code. As he put it:
> "Start a new agent under the same context, it remembers everything about your project. Share the context via link, anyone can continue building on the exact same shared context."
@iruletheworldmo described a lower-tech version that is arguably just as compelling. Give Codex a scratch pad file in your repo. Tell it to track its own mistakes, your corrections, and what worked. By session five, the agent is fixing things before you catch them. "Baby continual learning in a markdown file on my laptop," as they described it.
@blader validated the same intuition from yet another angle, noting that a "napkin" for agents to write on is meaningfully different from session history (which is lossy) or todos and plans (which are static). The running theme here is that agents need a lightweight, persistent, self-managed scratchspace. Not a database. Not a vector store. Just a file they can read and write across sessions. The simplicity is the feature.
AI Coding Tools Push Into the Terminal
GitHub made two related moves today. @_Evan_Boyle announced Copilot CLI is now available within VS Code, and @pierceboggan showed off the ability to spawn Copilot CLI terminals directly in VS Code Insiders. The terminal is becoming a first-class surface for AI assistance, not just the editor pane.
On the OpenAI side, @theo shared a video review after three weeks with Codex 5.3, calling it incredible while noting a few things he hopes OpenAI changes. And @RoxCodes described a workflow that feels like a glimpse of the near future: asking Codex to build something, then asking it to record a video testing the UI to prove it worked. It built a Playwright script, recorded the video, and attached it to the PR.
> "My #1 problem with ai coding is I never trust it to actually test stuff... the game changes every month now."
The trust gap in AI coding has always been verification. If agents can not only write code but also generate proof that the code works, that changes the adoption calculus for teams that have been skeptical about AI-generated PRs.
The "We're Still Early" Crowd Makes Its Case
@felixrieseberg from Anthropic offered a thought experiment that keeps circulating in AI circles: imagine models that are 10x faster, smarter, and more capable in specific domains, then imagine the products built on top of them improving by the same factor. He acknowledged this is useful even for skeptics, if only to understand "that weird 100-yard stare for 2026 some of us have."
The philosophical takes ranged from practical to poetic. @beffjezos distilled the long-term vision to one line: "The future is everyone owning the extension of their cognition, from the weights to the transistors." And @IterIntellectus went full contrarian, arguing the most valuable future skill is simply being someone people want to be around. Learn to cook, dance, make people laugh. When machines handle every automatable job, the economy becomes "how does it feel to be near you." It is a provocative frame, and whether or not you buy it completely, the underlying point lands: the skills that resist automation are fundamentally human ones.
Sources
One thing I’m missing from many takes on AI’s impact is that we’re still so, so early. I expect that we’ll continue to see dramatically better models, and even more dramatically better products on top of them. I believe the pace will surprise even many of those who are fully bought in and staying close to AI. Regardless of whether you think AI is overhyped or civilization-altering, I find it a useful exercise to imagine models that are 10x faster, smarter, and more capable in specific domains. Then, repeat the same exercise with the interfaces and products built on top of them. If nothing else, it’ll be super helpful in understanding the mindset of some of us at the frontier labs, even if you disagree with the speed we’re expecting. It’ll explain that weird 100-yard stare for 2026 some of us have.
Episode 10 of Raising An Agent with @sqs and @thorstenball is out! There's no better summary than this quote: "We will be killing our editor extension, the Amp VS Code extension. We're going to be killing it. And we're going to be killing it because we think it's no longer the future. We think the sidebar is dead. Let's walk through why." Topics in this episode: - The new deep mode in Amp - Balancing developer experience for humans & agents - Killing the VSCode extension & shift away from traditional editors - Pi & OpenClaw, two wonderful projects - Importance of reinventing yourself in AI Enjoy! And happy hacking! Timestamps: 01:00 Deep Mode 10:30 Optimizing the codebase for agents 15:00 Feature Preview: which Skills does your team use? 18:00 Balancing DX for humans & agents 21:35 Killing the Amp editor extension 28:00 The future of software and what it means 33:00 You need to stay agile 36:00 Pi & OpenClaw 39:00 Text editors holding companies back 44:00 Is manual context management coming to an end? 49:00 New concept for Threads 50:00 Amp, the business & the art installation
@jacobmparis @ScriptedAlchemy is creating the state of the art
LET ME COOK!! DGX Spark + Mac Studio + MBP + @exolabs + Mac mini M4 as my orchestrator. https://t.co/YwbVdCDLt6
Introducing OneContext. I built it for myself but now I can’t work without it, so it felt wrong not to share. OneContext is an Agent Self-Managed Context Layer across different sessions, devices, and coding agents (Codex / Claude Code). How it works: 1. Open Claude Code/Codex inside OneContext as usual, it automatically manages your context and history into a persistent context layer. 2. Start a new agent under the same context, it remembers everything about your project. 3. Share the context via link, anyone can continue building on the exact same shared context. Install with: npm i -g onecontext-ai And open with: onecontext Give it a try!