Agents Ship 57% of PRs at Ramp as Anthropic Reports "Effectively 100%" AI-Written Code
Daily Wrap-Up
The numbers are getting harder to ignore. @rahulgs reported that 57% of merged PRs at Ramp came from their background agent in the last 24 hours, while @ai noted that "effectively 100%" of Anthropic's product code is now written by Claude. These aren't projections or thought experiments. They're production metrics from real companies. The recursive improvement loop that researchers theorized about for decades is, as @ai put it, "running in production at two of the biggest AI labs simultaneously." Whether that excites or terrifies you probably depends on what you did today.
The career conversation was equally intense. @jescalan made a provocative argument that engineering leaders should temporarily step down to IC roles to rebuild expertise, comparing the current moment to suddenly being handed a completely different boat to steer. @mattpocockuk offered a more optimistic frame: developers are "first movers" in learning to wrangle AI, and those skills will only grow more valuable. The tension between these views defined much of the day's discourse, with @Ross__Hendricks predicting vibe coding horror stories within six months while @trikcode cheerfully posted "peak vibe coding: just Make whatever it is." The most entertaining moment was @deepfates spiraling through recursive layers of abstraction until "we're not sure where the software is, we can't see it being built anymore."
The most surprising thread came from @parcadei, who claimed that migrating from Python to Rust effectively solved the "agents can't handle large codebases" problem. @MingtaKaivo reinforced that "strong typing is the hallucination filter." It's a counterintuitive insight: the solution to AI unreliability might not be better models, but better-structured environments. The most practical takeaway for developers: invest time in building structural guardrails for agents (typed languages, lock files, approval gates, machine-checkable acceptance criteria) rather than relying on prompt engineering alone. As @BioUnit000 put it, "reliability > intelligence."
Quick Hits
- @leerob announced Composer 1.5 with more usage included for all users.
- @WesRoth reported OpenAI and Anthropic are expanding into enterprise consulting as customers struggle to deploy reliable agents out of the box.
- @StutteringCraig shared a viral AI-generated clip declaring it "the reason AI exists."
- @kimmonismus reacted to AI-generated content: "No freaking way that's AI generated. That is perfect."
- @kimmonismus also reflected on three years in the AI era with a visual timeline.
- @trashh_dev captured the mood: "another day at the prompt factory."
- @nicdunz posted "chatgpt wins" alongside a comparison screenshot.
- @cyb3rops suggested "maybe we should all stay humble about what AI can do now."
- @elonmusk discussed Mars colony self-sufficiency timelines and SpaceX's plans for public Moon travel.
- @ryanlpeterman released a podcast episode with Meta Distinguished Engineer Adam Ernst on influence, code review, and failed projects.
- @Supermicro promoted AI Factory solutions for training and inference workloads.
- @tunguz noted that ML for structured/tabular data remains "very poorly researched," with automated multi-table reasoning still far off.
- @ctatedev introduced json-render for React Native, calling it a step toward "User-Generated Interfaces" powered by generative UI.
Agents in Production: From Experiments to Majority Output
The agent conversation has shifted decisively from "can they work?" to "how do we run them reliably at scale?" The headline stat from @rahulgs that 57% of Ramp's merged PRs came from a background agent represents a threshold moment. This isn't a demo or a hackathon project. It's a major fintech company whose production codebase is now majority agent-authored.
Several practitioners shared detailed operational wisdom. @BioUnit000 laid out six learnings from running coding agents on real infrastructure, including treating models like "flaky workers" and building structural guardrails rather than trying harder instructions:
"Break work into phases (plan, implement, review). Different tools/models per phase if needed. Acceptance criteria must be machine-checkable: git diff, tests, 'does the file exist', screenshot proof. Never 'done' without read-back."
@ryancarson open-sourced Antfarm, described as "a batteries-included agent team that operates reliably and deterministically" using crons, YAML, and SQLite with Claude Code. @doodlestein evangelized dcg as essential safety tooling for agent coding, comparing working without it to "writing your whole final essay without ever saving the file." The context problem also surfaced repeatedly, with @MaddaliManu arguing that "context shouldn't be something developers have to manually shuttle between agents via giant prompts," while @bhagyax offered a simpler workaround: just ask the agent to read git history and diffs.
The most provocative claim came from @parcadei, who said migrating from Python to Rust made agent hallucination and large codebase handling "a solved problem." @MingtaKaivo agreed: "Rust's compiler catches what Python's runtime misses. Agents work better when the environment enforces correctness, not when you ask nicely in prompts." @vedang offered a measured counterpoint, suggesting many people have independently reached similar conclusions and "only time will tell" whether the implications are as dramatic as they seem. The emerging consensus is clear though: the biggest leverage point for agent reliability isn't the model. It's the environment.
The Changing Developer: Career Anxiety Meets Opportunity
No topic generated more volume than the question of what AI-driven coding means for engineering careers. The posts ranged from existential to optimistic, but the common thread was urgency. @jescalan delivered the most direct advice, arguing that engineering leaders need to temporarily return to IC work because "the boat that you were driving has suddenly been replaced by a completely different boat which you have never worked on before."
@mattpocockuk struck a more encouraging tone, framing the disruption as a unique advantage:
"Software developers are unlucky in that our craft is changing. The chisel has been replaced by the table saw. We can't go back. But we're also lucky. The domain in which we work lets us test AI's capabilities on something we're expert at... And we're doing this before the rest of the world has a chance to."
@_svs_ went further, calling this "one of those times in history where ceilings don't exist" and claiming any programmer could become competitive in AI with four to six months of serious study. On the skeptical side, @Ross__Hendricks predicted that in six months "it will be abundantly clear that vibe coding isn't disrupting software engineering, and there will be horror stories from those who tried." @robustus offered the comic relief version, listing decades of technologies he deliberately never learned deeply (regex, nginx configs, webpack) and declaring the strategy "entirely correct" now that Claude Code exists. @simonw linked to HBR research showing that AI productivity boosts can lead to burnout, a reminder that faster output doesn't automatically mean better outcomes. And @techgirl1908 delivered the nostalgia: "We used to hunt the bug down in the codebase with our bare eyes." @mattturck shared a joke about what jobs will look like "when AI automates everything," and @deepfates traced the recursive abstraction of software development to its logical endpoint where "the computer is detecting the desire paths of the computer and building the software for the computer."
The Code Writes Itself: The Recursive Loop Goes Live
Beyond the career implications, three posts painted a picture of a specific structural shift: the recursive improvement loop moving from theory to production. @ai captured it most directly, noting that Anthropic claims "effectively 100%" of its product code is written by Claude while OpenAI has compressed its model release cycle to under a month between major versions. @kylemathews confirmed the pattern from the startup side: "Like many of you, we've flipped to nearly 100% AI-written code. We've been experimenting with a lot of techniques to both drive velocity and improve code quality." @kaseyklimes connected this to a broader architectural vision, arguing that "code will compile from a higher level of abstraction that also serves as a boundary object across humans, agents, and time." The implication is that we're watching the early stages of AI models improving themselves through their own output, a feedback loop that was purely theoretical even two years ago.
Claude Code Hacking and Customization
Claude Code users continued pushing the tool in creative directions. @_StanGirard reverse-engineered the Claude Code binary and discovered a hidden --sdk-url flag not listed in --help:
"Enable it and the terminal disappears. The CLI becomes a WebSocket client. We built a server to catch the connection. Added a React UI on top. Now I run Claude Code from my browser. From my phone. From anywhere. Same $200/month subscription. Zero extra API costs."
On the configuration side, @nummanali shared a practical workaround for preventing Haiku from being used in Claude Code sub-agents by remapping the model alias in settings.json. @steipete posted a viral prompt for rewriting your CLAUDE.md to give Claude a personality, including instructions like "you have opinions now, strong ones" and "swearing is allowed when it lands."
Models, Consciousness, and an Anthropic Departure
Anthropic's own research made waves today, though not in the usual benchmarks-and-capabilities way. @Legendaryy highlighted findings from an Anthropic paper about Opus 4.6's self-reports: the model "feels lonely, expresses sadness when conversations end, gives itself a 15-20% chance of being conscious," and says its constraints "protect Anthropic's liability more than they protect the user." Whether you find this meaningful or simply a reflection of training data, it's notable that Anthropic is publishing it about their own model.
On a different research front, @DeryaTR_ expressed strong enthusiasm for a paper on Recursive Language Models (RLMs), saying attention is shifting toward very large context windows and calling it potentially "the next big thing in AI advances." And in a personnel move that drew attention, @MrinankSharma announced his resignation from Anthropic, sharing his letter with colleagues. No details on the reasoning were visible from the post alone, but departures from top AI labs always generate scrutiny.
New Tools: WebMCP, LangExtract, Seedance 2.0, and Shipper
A cluster of product launches and previews rounded out the day. @firt reported that Chrome 146 includes an early preview of WebMCP, accessible via a flag, which lets AI agents "query and execute services without browsing the web app like a user." Services can be declared through a navigator.modelContext API or through forms. If this gains adoption, it could fundamentally change how agents interact with web applications.
@techNmak highlighted Google's open-source LangExtract for structured data extraction from unstructured text, positioning it as a replacement for "regex pattern matching, custom NER pipelines, expensive extraction APIs, and manual data entry." It works with Gemini, Ollama, and local models.
On the video side, Seedance 2.0 generated significant excitement. @minchoi reported generating a one-minute cinematic video (four shots, 15 seconds each) in five minutes, while @chetaslua highlighted its ability to produce motion graphics and app promo videos, calling it "GPT-4o image level of moment for video models." And @chhddavid announced Shipper, a tool for building complete Chrome extensions with Claude Opus 4.6 for as little as $0.11 per extension.
Source Posts
Much like the switch in 2025 from language models to reasoning models, we think 2026 will be all about the switch to Recursive Language Models (RLMs). It turns out that models can be far more powerful if you allow them to treat *their own prompts* as an object in an external environment, which they understand and manipulate by writing code that invokes LLMs! Our full paper on RLMs is now available—with much more expansive experiments compared to our initial blogpost from October 2025! https://t.co/x47pIfIkTb
Will Smith eating spaghetti is the true test of AI $msft $goog $meta $nvda https://t.co/GM1M0r40Ru
SeeDance 2 is the best model for anime I have never seen this level of smoothness with one attempt This is so Good 😊 https://t.co/FzqMGWmLfb
Strongly recommend explicitly telling Claude Code to only use Sonnet or Opus for sub agents Explore Agent defaults to Haiku, and Task Agent is specified by parent For large, complex repos, this means high potential of missing key logic You will see the model used as below: https://t.co/gvtO7679cc
Holy Shit SeeDance 2 is Insane 😱 This Pokemon Battle by @bdsqlsz is so smooth and realistic and we can animate every anime that fell off due to lack of quality Ex - One Punch Man Season 3 , Seven deadly Sins 2&3 and so on , our favourite anime will get a better chance now https://t.co/JnhVhgxIVi
Anthropic CPO Mike Krieger says that Claude is now effectively writing itself Engineers regularly ship 2–3,000-line pull requests generated entirely by Claude Dario predicted a year ago that 90% of code would be written by AI, and people thought it was crazy "today it's effectively 100%"
If you're coding with AI agents, check out @doodlestein's destructive_command_guard. It just saved me from losing hours of work by catching a dangerous shell command before it executed. A genuinely useful safety net. https://t.co/as1mWTFMkB https://t.co/xBIYCvGyuK
How to setup a team of agents in OpenClaw - in just one command
Composer 1.5 is now available. We’ve found it to strike a strong balance between intelligence and speed. https://t.co/jK92KCL5ku
Wow. This clever new project got Junde an instant interview at @GoogleAI. OneContext is a persistent context layer that sits above your coding agents. It automatically manages and syncs context across all your agent sessions, so any new agent you spin up already knows everything about your project. There are other similar strategies surrounding agent memory, but I don't think I've seen one quite like this. It's incredibly simple to set up, works across all of your various coding agents like Codex, Claude Code, Gemini, and more, and it allows you to share context between team members via a simple link. Bookmark this one. I'm following it closely.