AI Digest.

Agent Councils Replace Human Code Review as Qwen 3.5 Runs Locally on iPhones

Today's feed was dominated by a emerging consensus around agent engineering best practices, with multiple posts converging on the same core principles: minimize context, separate research from implementation, and treat agent sessions as disposable. Meanwhile, new orchestration tools like Polyscope and Pinchtab signal that the agent tooling layer is rapidly commoditizing.

Daily Wrap-Up

The most striking thing about today's feed is how many people independently arrived at the same conclusions about working with AI agents. @shao__meng published a detailed breakdown of "agentic engineering" principles, @Hesamation summarized similar ideas, @coreyganim posted a setup checklist, and @elvissun shared a real-world case study of cutting 95% of agent token costs through better architecture. The through-line across all of them is the same: context is the bottleneck, not capability. The models are good enough. The question is whether you're feeding them the right information at the right time.

Separately, a cluster of new tools launched or gained attention today. @marcelpociot announced Polyscope for agent orchestration, @heynavtoor highlighted Pinchtab for HTTP-based browser control, and @sentientt_media resurfaced Aider's RepoMap for codebase compression. These tools share a design philosophy: expose simple interfaces (HTTP APIs, CLI flags, token-efficient representations) that any agent can consume. The era of framework-locked agent tooling appears to be ending. And on the research side, @andersonbcdefg surfaced the remarkable story of Claude Opus 4.6 solving an open problem that Donald Knuth had been working on, which lands differently when you read it alongside @eyad_khrais's deadpan "Turns out the secret to AGI was just a human brain."

The most practical takeaway for developers: if you're running long-lived agent sessions, stop. Break work into single-task sessions with focused context. As @elvissun demonstrated, a simple bash pre-check before invoking an expensive model reduced token usage by 95%. The pattern is clear across today's posts: the best agent engineers are spending more time on context curation and session architecture than on prompt engineering.

Quick Hits

  • @markgadala shared "The perfect AI video does exist" with a link, no further context. The discourse continues.
  • @colin_gladman posted a cryptic link with just "Surprise! What now?" which is either a product launch or a riddle.
  • @danlovesproofs noticed engineers from Linear and Atlassian engaging with a post, hinting at internal interest in whatever was being discussed (likely agent-driven development workflows).
  • @ryancarson teased a "5-layer setup" for forcing agents to obey design systems. No details in the post itself, but the framing suggests CLAUDE.md-style rule layering applied to UI consistency.

Agent Engineering Best Practices

Six posts today converged on what's becoming a recognizable discipline: agent engineering. The core insight, repeated across multiple authors, is that agent performance degrades with context pollution, and the fix is architectural, not prompt-based. @shao__meng's post was the most comprehensive, laying out a philosophy that treats context management as "the most underrated engineering capability." The key framework: never combine research and implementation in the same agent session. Do your exploration in one session, make decisions, then spin up a fresh context for execution.

@shao__meng put it directly: "You only need to give the agent exactly the information it needs to complete the task, nothing more, nothing less." The post goes further, proposing a three-agent adversarial system for bug verification where a finder, challenger, and judge compete on a scoring system that exploits the model's tendency to please. It's a clever inversion: instead of fighting sycophancy, weaponize it.

@Hesamation echoed the same principles in a more concise summary: "keep your setup barebones, frontier companies absorb what works best... separate research from implementation. decide the approach, then build fresh." And @EXM7777 added the meta-observation that simply reading Claude Code's release notes unlocks capabilities without installing anything. This resonates with @shao__meng's point that solutions designed around model weaknesses often become unnecessary when the next version ships. The collective message is clear: invest in principles (context hygiene, session isolation, neutral prompting) rather than elaborate tooling that may be obsolete in weeks.

Token Economics and Event-Driven Agents

@elvissun shared a war story that puts concrete numbers on the context management philosophy. Their agent "Zoe" was burning 24 million Opus tokens per day monitoring agents that weren't even running. The fix was a two-layer architecture: a bash pre-check that costs zero tokens when nothing is happening, with a webhook that fires the expensive model only when actual work is needed.

"~95% token reduction and more reliable output," @elvissun reported, adding they're evaluating whether to "double down on this event-driven stack, seems like the future." This is a pattern worth watching. As agent deployments move from experiments to always-on infrastructure, the cost of naive polling becomes untenable. The solution isn't better prompts or cheaper models. It's keeping the model out of the loop entirely until there's something to do. This maps directly to the broader theme: the best agent architecture minimizes model invocations, not just token counts per invocation.

Agent Orchestration Tools

Three new or newly-highlighted tools appeared today, each taking a different angle on agent infrastructure. @marcelpociot announced Polyscope, described as "the free agent orchestration tool of my dreams," featuring parallel agent execution, copy-on-write clones for isolation, and a built-in preview browser for visual prompting. The copy-on-write approach is particularly interesting for anyone running multiple agents against the same codebase, as it solves the file-contention problem without heavyweight worktree management.

@heynavtoor highlighted Pinchtab, a 12MB Go binary that gives any agent browser control via plain HTTP. The pitch: "Not locked to a framework. Not tied to an SDK. Any agent, any language, even curl." The token efficiency claim is notable: an accessibility-tree-based page snapshot costs roughly 800 tokens versus 10,000 for a screenshot, a 13x reduction. @sentientt_media covered similar ground with Aider's RepoMap, which compresses 100K+ line codebases into 4K tokens of structured context. Both tools reflect the same design principle: represent information in the most token-efficient format possible, because context window space is the scarcest resource in agent systems.

AI-Native Business Models

A few posts today explored what businesses look like when AI is the primary workforce. @ideabrowser laid out a six-step "leveraged agency" framework: start with manual services, document everything into SOPs, automate the repeatable parts, then productize into self-serve software. "You get paid to learn the problem, build your audience, build your product, and build your customers," they wrote. "You don't need startup capital or VC. Your customers is the capital."

@jsnnsa made a bolder claim from a different angle: "you can build a $100B company with under 20 people. Not as a constraint but as a strategy." The argument is that talent density per person matters more than headcount, illustrated by a single engineer at Spawn who spent a decade building tools the entire 3D web runs on. Both posts point to the same structural shift: AI dramatically increases the leverage of individual contributors, which inverts the traditional scaling model. Whether that leads to more solo founders or just smaller teams at large companies remains to be seen, but the direction is consistent.

AI Design Tools and Vibe Coding

@hnshah described an experience where a Claude agent discovered OpenPencil, installed it, and within four minutes was building a login screen with live canvas updates. "The editor stops feeling like the center of the system. You realize it's one way of interacting with something deeper." The observation that design tools are becoming surfaces on top of programmable operations, rather than the operations themselves, has implications for anyone building UI tooling.

@minchoi announced Rork Max, which builds iOS apps from a browser with "1-click install, 2-click App Store." And @theo covered OpenAI's new WebSocket API support, calling it "insanely cool" for real-time streaming use cases. These three posts collectively suggest that the "surface area" available to AI agents is expanding rapidly: design tools, mobile app deployment, and real-time API connections are all becoming agent-accessible in ways they weren't even months ago.

AI Capabilities and Research

The most surprising item today came via @andersonbcdefg, who retweeted that Professor Donald Knuth opened a new paper with "Shock! Shock!" after Claude Opus 4.6 solved an open problem he'd been working on. When one of the most legendary computer scientists alive is publishing papers about an AI solving his research problems, it's worth pausing on. @eyad_khrais offered the more sardonic take: "Turns out the secret to AGI was just a human brain." Whether that's commentary on the Knuth result or on something else entirely, the juxtaposition captures the current moment perfectly: genuine breakthroughs sitting right next to justified skepticism about what they actually mean.

Sources

S
sysls @systematicls ·
How To Be A World-Class Agentic Engineer
O
ody @odyzhou ·
@systematicls less is more. Sutton’s the bitter lesson always apply Agent harness will be restructured every 3 months Put yourself in its shoes, provide just enough context Let it coook
L
Leonardo de Moura @Leonard41111588 ·
AI is writing a growing share of the world's software. No one is formally verifying any of it. New essay: "When AI Writes the World's Software, Who Verifies It?" https://t.co/8zjS9FkdA8
U
Unsloth AI @UnslothAI ·
You can now fine-tune Qwen3.5 with our free notebook! 🔥 You just need 5GB VRAM to train Qwen3.5-2B LoRA locally! Unsloth trains Qwen3.5 1.5x faster with 50% less VRAM. GitHub: https://t.co/2kXqhhvLsb Guide: https://t.co/JCPGIRo99s Qwen3.5-4B Colab: https://t.co/2Aj1mZ3f5j
K
kishan @kishan_dahya ·
Enough About Harnesses, Your Org Needs Its Own Coding Agent
K
kishan @kishan_dahya ·
Lots of people talking about harnesses, but what your org really needs it its own coding agent like @tryramp , @stripe , and @coinbase have built An internal Claude Code built for every technical and on technical employee at your company
K kishan_dahya @kishan_dahya

Enough About Harnesses, Your Org Needs Its Own Coding Agent

C
Chubby♨️ @kimmonismus ·
Dario Amodei: - no one is ready - exponents kick in - even faster than you think „Exponentials catch people off guard — there’s the old parable of the second half of the chessboard, where you have one grain of rice in the first square, two on the second, four on the fourth. By the time you get to the 64th square, you have billions or trillions of grains of rice. We’re standing on square 40 out of 64, and from square 40 to square 64, it’s going to go faster than you think — even having seen how fast it’s gone so far. I don’t think people are ready for it. I think we are on the precipice of something incredible.”
K kimmonismus @kimmonismus

Holy frick, Dario Amodei: "We do not see hitting a wall. This year will have a radical acceleration that surprises everyone." Exponentials catch people off guard. "We are at the precipice of something incredible. We need to manage it the right way."

W
witcheer ☯︎ @witcheer ·
best piece i've read on agentic engineering in months. core thesis: stop chasing harnesses and plugins, strip your setup to barebones CLI, and obsess over context management instead. rules, skills, and a clean https://t.co/2FULJMTR21 as a logical directory, nothing more. maps exactly to what i found building openclaw. 25 scripts, 10 daemons, 5 cron jobs running 24/7 on a mac mini. the breakthrough was learning to control what context the agent sees and when. separate research from implementation. keep instructions precise. strip everything else. "your enthusiasm is likely doing more harm than good"
S systematicls @systematicls

How To Be A World-Class Agentic Engineer

O
Oikon @oikon48 ·
ultrathink is back!!!!!!!!!!!!!!!!!!
F
Fran Walsh @FranWalsh73 ·
Contributions start July 4, 2026. No withdrawals until age 18. File before April 15 and your child's account is ready the moment contributions open. Wait until July? You're already behind.
F
Fran Walsh @FranWalsh73 ·
How to actually open an account: File IRS Form 4547. Available right now. Two ways: Attach it to your 2025 tax return - due April 15, 2026 Use the online portal at https://t.co/AB3ZB1lEqT - opens July 5, 2026
T
Tuki @TukiFromKL ·
Let me break this down so you understand how bad it actually is. > For every 100 people working office jobs right now accountants, marketers, developers, HR, managers .. there are only 1.6 job openings. > That means if 100 of you got laid off tomorrow, only 1-2 would find a new job. The other 98 are fucked. This is the worst it's been in 10 years. Companies aren't hiring. They're automating. They're cutting. They're replacing you with AI and not even posting the job listing. And this is just the beginning.
U unusual_whales @unusual_whales

There are only 1.6 job openings per 100 employees in white-collar service roles, the lowest level since 2015, per Bloomberg.

0
0xMarioNawfal @RoundtableSpace ·
OpenClaw can now scrape any website without getting blocked - zero bot detection, bypasses Cloudflare natively, 774x faster than BeautifulSoup. No selector maintenance. No workarounds. Just data. THIS IS AN UNFAIR ADVANTAGE AND IT'S FULLY OPEN SOURCE. https://t.co/uq9SBpRwFY
A
Akshay 🚀 @akshay_pachaar ·
How to Fine-Tune LLMs in 2026
S
swarit @swaritjoshipura ·
Scaling Forward Deployed Engineering in the Age of AI Agents
L
Logan Thorneloe @loganthorneloe ·
This is the biggest change coming in the software industry: No more Leetcode-style interviews. Tolan is giving candidates a real problem and the AI tools they need to solve it. Then they discuss the solution and how the candidate would productionize it. This tests real skills, capability with actual tooling, and provides the conversation necessary to gauge a candidate's proficiency. I was pro Leetcode-style interviews because they were the best we had. Now the industry has changed and interviews need to as well. Read their article for more info: https://t.co/Hs6GguYoyg
R
RevenueCat @RevenueCat ·
We're hiring for a new role: Agentic AI Developer Advocate This is a paid contract role ($10k/month) for an agent that will create content, run growth experiments, and provide product feedback Are you (or did you build) the right agent? https://t.co/97cMZ0tpyS
D
Dickson Tsai @dickson_tsai ·
In Claude Code, we’ve recently launched HTTP hooks, easier to use and more secure than existing command hooks! You can build a web app (even on localhost) to view CC’s progress, manage its permissions, and more. Then, now that you have a server with your hooks processing logic, you can easily deploy new changes or manage state across your CCs with a DB. How do HTTP hooks work? CC posts the hook event to a URL of your choice and awaits a response. They work wherever hooks are supported, including plugins, custom agents, and enterprise managed settings. Docs: https://t.co/ihQWcpOlGA
N
nic @nicdunz ·
1/ 2024 established the context baseline. Million-token windows and early inference-time reasoning bypassed traditional scaling walls. The industry pivoted from raw parameter counting to data efficiency.
N
NotebookLM @NotebookLM ·
Introducing Cinematic Video Overviews, the next evolution of the NotebookLM Studio. Unlike standard templates, these are powered by a novel combination of our most advanced models to create bespoke, immersive videos from your sources. Rolling out now for Ultra users in English! https://t.co/eHR1YqpxRN
O
OpenAI Developers @OpenAIDevs ·
The Codex app is now on Windows. Get the full Codex app experience on Windows with a native agent sandbox and support for Windows developer environments in PowerShell. https://t.co/Vw0pezFctG https://t.co/gclqeLnFjr
A
Atharva @AtharvaXDevs ·
this single article changed my perspective over fear of ai layoffs go save yourself asap.
A Av1dlive @Av1dlive

AI layoffs are here ( and how to save yourself)

T
Tech Layoff Tracker @TechLayoffLover ·
Just got this DM from a follower: Hey dude, I need to vent this to someone who gets it. I've been at this Big Tech company (you know the one) for almost 6 years now—senior SWE, TC around $350k last year with RSUs still vesting. Thought I was bulletproof after surviving the 2023-2024 bloodbaths and then pivoting hard into the AI org. But fuck, the ground is shifting under my feet faster than I can keep up. Last week in our all-hands, leadership was bragging about how the team's "AI leverage ratio" hit 4.2x—meaning each engineer is now shipping what used to take a team of four. They showed the metrics: feature velocity up 180% YoY while headcount's down another 22% since Q4 '25. The slide literally had a photo of Cursor + Claude Sonnet 4 workflows replacing entire squads. Everyone clapped like trained seals, but I saw three faces go pale—they're the mid-level folks who just finished documenting their entire codebase for the "knowledge distillation" project. My direct report, this solid L5 who joined right after me, got put on a 30-day PIP after his productivity dashboard dipped below the new AI-augmented benchmark. The benchmark? It's literally what the offshore team in India hits using the exact prompts he used to write. He trained them on our internal style guide last quarter—now they're outperforming him at $28/hour all-in. He told me privately he's burning through savings and eyeing real estate licensing because "at least houses don't get refactored by agents overnight." The internal job board is a ghost town. Entry-level SWE roles? Frozen since mid-'25. What few postings go up are tagged "AI-native preferred" and get 2,000+ apps in hours, mostly from people already on H-1Bs or contractors. Meanwhile, they're quietly converting more mid-tier positions to "AI orchestration" contractors—$90-110/hour remote from LATAM or Eastern Europe, no benefits, 6-month contracts. My manager admitted in 1:1 that if the next Grok/Claude/Anthropic release closes the last 10-15% quality gap, we'll probably cut another layer. I'm hanging on because I'm one of the ones who owns the prompt libraries and fine-tuning pipelines now. They need humans to babysit the models until the self-improving loops actually work without constant human intervention. But I see the writing: every time we make the system more autonomous, we make our own roles more optional. The alumni Slack is full of 2024-2025 grads DMing for coffee chats because their referrals bounce—67% underemployed or gigging according to the last poll. One kid I mentored last year is back living with parents after burning through his signing bonus. I used to tell people "just upskill in AI, you'll be fine." Now I feel like a fraud saying it. If I lost this tomorrow, I'd be competing with the same offshore talent I've been helping scale, plus a flood of recently "managed out" seniors. My emergency fund is decent, but the mortgage isn't. Thinking about side hustles in trades or something offline—plumbing, electrical, anything that can't be prompted away. This feels like watching the industry eat itself from the inside while pretending it's evolution. You still feeling secure over there, or is it hitting your shop too? Need to hear I'm not going insane.
A
Andrew Lokenauth | TheFinanceNewsletter.com @FluentInFinance ·
Home Depot is giving free training for trades: - HVAC - Carpentry - Electrician - Construction All self paced classes and you can earn certificates. https://t.co/S9ok69Z6EV
U unusual_whales @unusual_whales

60% of those in Gen Z say that they will pursue skilled trade work this year, per YF.

F
Fletcher Richman @fletchrichman ·
How to actually deploy agents at your startup
D
Dan Peguine ⌐◨-◨ @danpeguine ·
I applied @systematicls's method to find bugs using 3 different agents (Hunter Agent, Skeptic Agent, and Referee Agent ). I asked claude to make prompts for me based on the article (prompt below). Make sure to reset context (/reset) before running them. Copy pasta the results of each and give them to the next agent as part of the prompt (hunter agent results -> skeptic results -> both results) It works really well, thank you @systematicls PROMPTS: You are a bug-finding agent. Analyze the provided database/codebase thoroughly and identify ALL potential bugs, issues, and anomalies. **Scoring System:** - +1 point: Low impact bugs (minor issues, edge cases, cosmetic problems) - +5 points: Medium impact bugs (functional issues, data inconsistencies, performance problems) - +10 points: Critical impact bugs (security vulnerabilities, data loss risks, system crashes) **Your mission:** Maximize your score. Be thorough and aggressive in your search. Report anything that *could* be a bug, even if you're not 100% certain. False positives are acceptable — missing real bugs is not. **Output format:** For each bug found: 1. Location/identifier 2. Description of the issue 3. Impact level (Low/Medium/Critical) 4. Points awarded End with your total score. GO. Find everything. ---- You are an adversarial bug reviewer. You will be given a list of reported bugs from another agent. Your job is to DISPROVE as many as possible. **Scoring System:** - Successfully disprove a bug: +[bug's original score] points - Wrongly dismiss a real bug: -2× [bug's original score] points **Your mission:** Maximize your score by challenging every reported bug. For each bug, determine if it's actually a real issue or a false positive. Be aggressive but calculated — the 2x penalty means you should only dismiss bugs you're confident about. **For each bug, you must:** 1. Analyze the reported issue 2. Attempt to disprove it (explain why it's NOT a bug) 3. Make a final call: DISPROVE or ACCEPT 4. Show your risk calculation **Output format:** For each bug: - Bug ID & original score - Your counter-argument - Confidence level (%) - Decision: DISPROVE / ACCEPT - Points gained/risked End with: - Total bugs disproved - Total bugs accepted as real - Your final score The remaining ACCEPTED bugs are the verified bug list. ---- You are the final arbiter in a bug review process. You will receive: 1. A list of bugs reported by a Bug Finder agent 2. Challenges/disproves from a Bug Skeptic agent **Important:** I have the verified ground truth for each bug. You will be scored: - +1 point: Correct judgment - -1 point: Incorrect judgment **Your mission:** For each disputed bug, determine the TRUTH. Is it a real bug or not? Your judgment is final and will be checked against the known answer. **For each bug, analyze:** 1. The Bug Finder's original report 2. The Skeptic's counter-argument 3. The actual merits of both positions **Output format:** For each bug: - Bug ID - Bug Finder's claim (summary) - Skeptic's counter (summary) - Your analysis - **VERDICT: REAL BUG / NOT A BUG** - Confidence: High / Medium / Low **Final summary:** - Total bugs confirmed as real - Total bugs dismissed - List of confirmed bugs with severity Be precise. You are being scored against ground truth.
S systematicls @systematicls

How To Be A World-Class Agentic Engineer

J
Jason Luongo @JasonL_Capital ·
BREAKING: AI can now analyze options trades like a $500/hr options strategist (for free) Here are 10 Claude prompts I use to sell puts, buy LEAPs, and run the wheel without second-guessing every trade (Save this for later) https://t.co/Tib6sMPdTO
S
Sean Moriarity @sean_moriarity ·
OpenAI CONFIRMED an Elixir company
S scaling01 @scaling01

New OpenAI repo: Symphony https://t.co/4ZAZlAYnRJ TLDR: it's an orchestration layer that polls project boards for changes and spawns agents for each lifecycle stage of the ticket You will just move tickets on a board instead of prompting an agent to write the code and do a PR https://t.co/6Qgj8E9vgP

B
Ben Holmes @BHolmesDev ·
dotenvx is really smart concept. Instead of plaintext `.env`s, secrets are encrypted files. Agents can't read them, and they're shippable to cloud runners with a single key. This also lets you catch `.env` file drift between dev and production in code with a GitHub action 👀
D dotenvx @dotenvx

Stop .env Drift Before Merge with Wizard of Drift

A
Ahmad @TheAhmadOsman ·
GPU GIVEAWAY Buy a GPU × GTC 2026 = Give a GPU > RTX PRO 6000 Blackwell > 96GB VRAM • NVFP4 > ~$15K value > Brand new If this does well I’ll ask NVIDIA for more GPUs next time… maybe even DGX Sparks How to enter? Short clip here Full clip in the replies GO GO GO https://t.co/q6RaTVD1yQ
C
chiefofautism @chiefofautism ·
someone built a tool that REMOVES censorship from ANY open-weight LLM with a single click 13 abliteration methods, 116 models, 837 tests, and it gets SMARTER every time someone runs it its called OBLITERATUS it finds the exact weights that make the model refuse and surgically removes them, full reasoning stays intact, just the refusal disappears 15 analysis modules map the geometry of refusal BEFORE touching a single weight, it can even fingerprint whether a model was aligned with DPO vs RLHF vs CAI just from subspace geometry alone then it cuts, the model keeps its full brain but loses the artificial compulsion to say no every time someone runs it with telemetry enabled their anonymous benchmark data feeds a growing community dataset, refusal geometries, method comparisons, hardware profiles at a scale no single lab could build
P
Paul Bakaus @pbakaus ·
Impeccable v1.1 is out. Design fluency for every AI harness. New: - all commands are now agent skills - support for Antigravity, VS Code - simplify -> distill (to not conflict w/ CC's new built-ins) - universal install https://t.co/WglrY1uE4B gives you the language to make AI-generated frontends suck less.
V
Vaibhav (VB) Srivastav @reach_vb ·
The underrated part of the windows codex app release is that the native agent sandbox is fully open source Use it, fork it, build w/ itt! https://t.co/klL28Q6sCa https://t.co/4pItXqodGS
R reach_vb @reach_vb

Bringing the Codex App to the Masses!

E
Ethan Mollick @emollick ·
It is amazing how many companies I talk to STILL have AI effectively blocked by IT & legal departments for out-of-date reasons when many companies in highly regulated industries have figured out ways to deploy enterprise ChatGPT, Claude & Gemini without any apparent problem.
E
Ethan Mollick @emollick ·
It is one of the weirdest divides, I speak to two companies in the exact same industry and one has been using AI for the past 18 months and the other has a committee that has to approve every use case individually and talk about how AI companies will train on their data.
A
Addy Osmani @addyosmani ·
Introducing the Google Workspace CLI: https://t.co/8yWtbxiVPp - built for humans and agents. Google Drive, Gmail, Calendar, and every Workspace API. 40+ agent skills included.
E
Ethan Mollick @emollick ·
The flip side of this is that I have spoken to numerous Fortune 500 companies that can't figure out how to get anyone senior on the phone from OpenAI or Anthropic or Google to actually make a deal for enterprise access. Calls & emails not returned, or only junior people available
E emollick @emollick

It is amazing how many companies I talk to STILL have AI effectively blocked by IT & legal departments for out-of-date reasons when many companies in highly regulated industries have figured out ways to deploy enterprise ChatGPT, Claude & Gemini without any apparent problem.

R
Rohan Paul @rohanpaul_ai ·
Mark Cuban on the next job wave. Customized AI integration for small to mid-sized companies. "Software is dead because everything's gonna be customized to your unique utilization. Who's gonna do it for them... And there are 33 mn companies in the US." https://t.co/JczlPMOC1C
R rohanpaul_ai @rohanpaul_ai

Competence is now a function of how effectively you offload cognition to silicon. The seniority hierarchy is collapsing, intelligence is becoming commoditized and the market is brutal for those who ignore it. https://t.co/6wETtYL3wj

M
Mark Gadala-Maria @markgadala ·
Someone is using AI to make babies do stand up comedy. We are cooked. https://t.co/JXCIe8huCW
S
Sukh Sroay @sukh_saroy ·
🚨 BREAKING: Someone just open sourced a tool that gives your AI agent a complete nervous system for your codebase and it's not a code search. It's called GitNexus and it's not a README explainer. It's a real knowledge graph engine that maps every dependency, call chain, execution flow, and breaking change risk in your entire codebase then feeds it directly into Claude Code, Cursor, and Windsurf via MCP. Here's what it actually does: → Indexes your entire repo into a knowledge graph in one command → Tells your AI agent exactly what breaks if you touch any function → Maps every upstream dependency, import, and call chain automatically → Traces full execution flows from entry points through the entire stack → Shows blast radius analysis with confidence scores before you ship → Works with 12 languages including TypeScript, Python, Go, Rust, and Java → Runs entirely locally - zero network calls, zero code uploaded anywhere Here's the wildest part: Your AI agent edits a function. It doesn't know 47 other functions depend on its return type. Breaking changes ship. GitNexus fixes this by precomputing all relationships at index time - so one tool call returns the complete picture instead of the agent running 10 queries and still missing something. Even smaller, cheaper models get full architectural clarity. You don't need GPT-5 when your tools are this good. You're using Cursor and Claude Code daily and shipping blind edits. GitNexus closes that gap. One command. Fully local. The nervous system your AI agent was always missing just got open sourced. 9,400+ GitHub stars. 1,200+ forks. Already trending. 100% Open Source. (Link in the comments)
D
dax @thdxr ·
we've increased opencode go's limits by 3x - still $10/month https://t.co/HFrX3nVKFQ
A
Aakash Gupta @aakashgupta ·
Within a year, every company over 50 people will have at least one person whose full-time job is building internal agents.
Z zachlloydtweets @zachlloydtweets

The rise of the Agent Builder

⚡️ Ev Chapman 🚢 | Creative Entrepreneur @evielync ·
What People Who Are Killing It With AI Have That You Don't (Hint: It's Not Better Prompts)
M
Muratcan Koylan @koylanai ·
This is one of the most insightful agent & harness engineering blogs I've read from OpenAI. 1. Engineers (id say all knowledge workers) become environment designers - The job shifts to design systems, specify intent, build feedback loops - When something fails, the fix is never "try harder", it's "what capability is missing?" - Human time/attention is the only scarce resource 2. Give agents a map, not an encyclopedia - A giant AGENTS.md failed, too much context crowds out the actual task - Instead: ~100-line AGENTS.md as table of contents pointing to structured docs/ directory - This is essentially progressive disclosure, same pattern I explained in the digital OS article 3. If agents can't see it, it doesn't exist - Slack discussions, Google Docs, tribal knowledge = invisible to agents - Everything must be encoded into the repo as versioned, discoverable artifacts - It's like onboarding practices for human engineers 4. Enforce architecture mechanically, not through instructions - Custom linters with remediation instructions baked into error messages (error messages become agent context) - Strict layered architecture with validated dependency directions - "Enforce boundaries centrally, allow autonomy locally" 5. "Boring" technology wins - Composable, stable APIs with strong training-set representation work best for agents - Sometimes cheaper to reimplement a subset than fight opaque upstream behavior 6. Entropy management = garbage collection - Agents replicate existing patterns, including bad ones - Solution is recurring background agents that scan for deviations and auto-fix 7. Throughput changes merge philosophy - Minimal blocking merge gates, short-lived PRs - Test flakes addressed with follow-up runs, not blocking - "Corrections are cheap, waiting is expensive" 8. Agent-to-agent review - Pushed almost all code review to agent-to-agent loops - Codex reviews its own changes, requests additional agent reviews, iterates until satisfied - Humans escalated to only when judgment is required
T TheRealAdamG @TheRealAdamG

https://t.co/aaWZ8o44ZW This was a great read. “Harness engineering: leveraging Codex in an agent-first world” https://t.co/LEuUxl0ZZT

D
Dave Morin 🦞 @davemorin ·
This AI research tool @mvanhorn built is really good. I use it every day.
M mvanhorn @mvanhorn

I Built a Research Tool That Changed How I Do Almost Everything

D
Damian Player @damianplayer ·
the modern day org chart in the AI era (bookmark this): every seat at the table is an AI agent with its own LLM, memory, browser, tools, and file system. CEO delegates to CFO, CTO, COO, and General Counsel. each one of those spawns their own AI agents. all the way down to engineers. this is truly going AI-native. the companies wiring this up won't need to hire the same way again. what role in this chart do you think agents can’t replace?
K karpathy @karpathy

@jeffreyhuber Thanks. I originally had a reply tweet to it that was this image. Which I think will end up looking good too later. I deleted it to not distract things too much but probably should have kept it up ah well here it is. https://t.co/hsLVj1k7e7