Ramp's Inspect Agent Authors 30% of Merged PRs While Ralph Wiggum Tooling Ecosystem Proliferates
Ramp shared hard numbers on their internal coding agent Inspect, now responsible for 30% of merged PRs with non-engineers submitting code. The Ralph Wiggum agentic workflow pattern spawned multiple competing CLI tools in a single day. Claude Code continued expanding into non-engineering workflows as developers debated whether senior expertise matters more than AI-assisted speed.
Daily Wrap-Up
The most concrete signal today came from Ramp, where @eglyman dropped real production numbers: their internal coding agent Inspect now authors roughly 30% of merged PRs across core repos, and people from "essentially every job function" submitted code last week. That's not a demo or a hackathon stat. That's a Fortune 500 fintech running a significant chunk of its engineering output through an autonomous agent with feedback loops built around tests, telemetry, and visual checks. @thdxr's reaction captured the vibe perfectly: "it's insane how much ramp built." The gap between companies that have figured out agent-assisted development and those still debating whether to adopt Copilot is widening fast.
On the tooling side, the Ralph Wiggum pattern of agentic task orchestration is clearly hitting escape velocity. Two independent developers shipped competing CLI tools on the same day: @iannuttall built a multi-agent ralph CLI that works across Codex, Claude, and Droid, while @theplgeek dropped ralph-tui with plugin architecture and end-to-end observability. When multiple people independently build tooling around the same workflow pattern, it's a strong signal that the pattern has legs. The agentic loop of PRD generation, plan decomposition, and autonomous execution is becoming a standard workflow rather than an experiment.
The career anxiety thread ran hot as usual, with @ujjwalscript's story about a junior dev outperforming a 10-year veteran using three orchestrated AI agents landing as the day's most polarizing post. His conclusion that senior developers should "compete on wisdom, not speed" is comforting but incomplete. The more honest read is that architectural judgment still matters enormously, but the window where that judgment alone justifies a senior title is narrowing. The most practical takeaway for developers: study how Ramp built Inspect's feedback loops (tests, telemetry, feature flags, visual checks). The winning pattern isn't "AI writes code" but "AI writes code and then verifies it against reality." Build that verification layer into your own workflows now, whether you're using Claude Code, Cursor, or a custom agent setup.
Quick Hits
- @steveruizok dropped a cryptic "$100B app" post with no further context. Filing under "things that will make sense in a week or never."
- @doodlestein pointed someone to a new project replacing their previous work, linking to a repo without much explanation.
- @steipete released bird 0.7.0, a fast X CLI for reading tweets, now with home timeline support, trending/news, and user-tweets. Multiple community contributors credited.
Claude Code Expands Beyond Engineering
The most interesting Claude Code development today wasn't about writing better code. It was about not writing code at all. @boringmarketer declared "Claude Code for non technical work is massive," pointing to a trend that's been building for weeks: the tool's filesystem-based workflow, context management, and iterative execution model work just as well for non-engineering tasks as they do for shipping features. @rohanvarma fleshed this out with a detailed example from a Fortune 500 company where PMs are using Cursor (same general category) to run their entire product workflow in code:
> "A GitHub repository for all PMs. Customer call transcripts checked directly into the repo. Cursor agents extract insights from those transcripts and write them to a dedicated insights directory. PRDs are generated into a separate folder, creating a durable record of product decisions that agents can reference later."
This is "PMing in code," as @rohanvarma put it, treating product work as "an evolving, inspectable system rather than a collection of docs and meetings." The implication is significant: once your workflow lives in a repo with rules files and agent instructions, you get version history, diffing, branching, and automation for free. That's a better knowledge management system than most companies have ever had.
On the technical side, @parcadei highlighted the token cost problem that anyone running agents at scale has hit: paying 320K tokens to read a single file when a smarter summarization approach costs 400. Context window management is becoming a core competency for agent builders, not just a nice-to-have optimization. And @DanielMiessler flagged a category of software that's suddenly vulnerable: mid-quality, niche tools that won because they were the only option. "Claude Code can just reverse engineer it," he noted, which is both a threat to incumbents and an opportunity for anyone who's been stuck using mediocre tooling. @eyad_khrais continued building out educational content with a level 2 Claude Code tutorial, feeding the growing demand for structured learning around these tools.
Agents Hit Production: Ramp's Inspect and the Feedback Loop Pattern
Ramp's @eglyman provided what might be the most important data point of the day for anyone building or evaluating coding agents. Their Inspect agent doesn't just generate diffs. It runs in sandboxed dev environments, observes test results, checks telemetry, evaluates feature flags, and for UI work, takes screenshots and compares live previews. The key insight is framing agents as control systems:
> "Generating output is easy. Feedback is everything... It doesn't just propose diffs; it iterates until the evidence says the change is correct."
Two specific observations stood out. First, cheap parallel sessions change behavior fundamentally. When the agent runs in its own sandbox rather than on your laptop, you stop micromanaging and start running more experiments. Second, the multiplayer dimension matters more than expected. Inspect shows up in PRs, Slack, VS Code, and the web, and sessions can be handed off between teammates. It becomes "shared infrastructure, not a novelty."
The 30% merged PR stat is striking, but the detail that non-engineers are submitting code is arguably more significant. That's the bridge between "AI helps developers code faster" and "AI changes who can contribute to a codebase." @thdxr's reaction captured the developer community's mix of admiration and anxiety: Ramp essentially built out someone else's entire product roadmap as an internal tool. The build-vs-buy calculus is shifting when your internal agents can ship features this fast.
The Ralph Wiggum Tooling Boom
Two independent Ralph Wiggum CLI tools shipped on the same day, which says something about where the agentic workflow community is headed. @iannuttall built a cross-platform ralph CLI that works with Codex, Claude, and Droid, combining PRD creation with plan generation and a ralph build command to execute:
> "works with codex, claude, droid. creates a prd for you. turns prd into a plan. run ralph build to cook"
Meanwhile, @theplgeek went deeper with ralph-tui, a terminal UI that adds plugin architecture for both agents and task trackers, built-in PRD creation leveraging skills, task dependency understanding, and what they describe as "end-to-end observability." The plugin system ships with Claude Code and OpenCode integrations out of the box, plus JSON and Beads tracker formats.
The convergence is notable: both tools independently arrived at the same core workflow of PRD-to-plan-to-execution, but took different approaches to extensibility. @iannuttall optimized for simplicity and multi-agent support, while @theplgeek built for observability and plugin ecosystems. This kind of parallel evolution typically means the underlying pattern is solid and the community is entering a differentiation phase where tooling competes on developer experience rather than core capabilities.
Local AI: From Prediction to Practice
@TheAhmadOsman showed up twice today, once with predictions and once with receipts. The prediction thread was straightforward: open-source AI wins, AGI runs local, learn how it works now. The receipts were more interesting: a working demo of Claude Code running against local models served by vLLM on 4x RTX 3090s, with GLM-4.5 Air handling generation:
> "vLLM serving GLM-4.5 Air on 4x RTX 3090s. nvtop showing live GPU load. Claude Code generating code + docs. end-to-end on my AI cluster. this is what local AI actually looks like"
The practical reality of local inference is still rough around the edges. Consumer GPU setups require significant investment and expertise, and the models you can run locally trail the frontier by meaningful margins. But the gap is closing, and for certain workflows, especially ones involving proprietary codebases or compliance requirements, having the entire stack on your own hardware is worth the tradeoffs. The "buy a GPU" advice is becoming less meme and more genuine career investment for developers who want to understand the full stack of AI-assisted development.
AI and the Senior Developer Identity Crisis
@ujjwalscript's post about a junior developer outperforming a 10-year veteran hit a nerve, as these posts always do. The story follows a familiar arc: junior uses three orchestrated AI agents to ship in 4 hours what would take 3 days, but the PR has security holes and lacks architectural vision. The conclusion lands on "compete on wisdom, not speed," which is reassuring but glosses over the uncomfortable middle ground where "wisdom" needs to be encoded into agent instructions and review processes rather than held in someone's head.
@DaveShapi took a much longer view with a thread on preparing for AI job displacement, covering everything from relocating to lower cost-of-living areas to identifying remaining job categories (attention economy, experience economy, authenticity economy, meaning economy). It's a sobering framework, though the practical advice of "save more, invest wisely, find purpose beyond work" applies regardless of whether AI displaces your specific role. The more actionable signal for working developers is in the Ramp data: the companies moving fastest aren't replacing engineers. They're augmenting everyone into a contributor. The question isn't whether your job disappears but whether you're building the systems that make that augmentation work, or waiting for someone else to build them around you.
Sources
I replicated a $5K scroll animation inside Cursor in 10 minutes. People keep saying AI can’t replace designers. That might be true for big companies with huge teams and complex design systems. But if your goal is to ship an MVP fast, Gemini 3 or Opus 4.5 is MORE than enough. I one-shotted a landing page with a scroll animation agencies charge thousands for. Here’s the exact process I used ↓
how the creator of claude code actually writes software
the creator of claude code just revealed his personal setup and it makes every other workflow look obsolete. Boris Cherny runs 5 Claude instances in h...
Unit makes Metaprogramming trivial. I can quickly turn this web server into a *Hot Web Server*: Every change made to the website's source is immediately propagated to all users, no reload nor reinstall needed. Imagine being able to solve your users problems... immediately. ⚡️ https://t.co/U3ZEMbHDU4
How to Make Realistic Longform AI Videos (Prompts Included)
This is going to be a step by step breakdown on how to make longform AI videos that LOOK and SOUND realistic… So you can push out crazy amounts of rea...
millennial gamers are the best prepared generation for agentic work, they've been training for 25 years https://t.co/JHsbPQHupk
I almost quit Codex after 1 day. Here's how to actually use it.
I almost rage-quit Codex after one day. It doesn't infer intent as well as Claude Code. I literally pasted in an error log and it said "what do you wa...
Here's what we've learned from building and using coding agents. https://t.co/PuBtYuhyhd
How I Use Claude Code
I built one of our most complex features - learning machines - in 5 days. 100% of the code was written by claude code. This would've taken months befo...
Collaborative Intelligence
I used Claude Code to reverse-engineer the Claude macOS Electron app and had Cowork dig around in its own environment - now I've got a good idea of how the sandbox works It's an Ubuntu VM using Apple's Virtualization framework, details here: https://t.co/lRWVhrNFk0