AI Learning Digest.

Claude Code Ships Task Management and Swarm Mode as Skills Ecosystem Reaches Critical Mass

Daily Wrap-Up

Today marked a clear inflection point for Claude Code's autonomy story. Anthropic shipped the upgrade from Todos to Tasks, and with it came multi-agent swarm spawning directly from plan mode. That alone would have been the headline, but the real story is how quickly the surrounding ecosystem responded. @rauchg confirmed that skills adoption exceeded his expectations, @supabase launched Postgres best-practices skills, and @vercel_dev made frontend design skills a single npx command away. The convergence of structured task management, swarm orchestration, and a growing skills marketplace means Claude Code is rapidly becoming less of a coding assistant and more of a development platform. Meanwhile, GitHub and Microsoft made their countermove by releasing the Copilot SDK, letting developers embed the same agentic runtime behind Copilot CLI into any application.

Beyond the tooling wars, the agent runtime space is fragmenting in fascinating ways. @martin_casado from a16z endorsed the Sprite model of persistent Linux environments running AI agents with checkpoints instead of git. @irl_danB counted four new "intelligent VM" attempts in three weeks. The pattern is clear: the industry is converging on the idea that agents need full compute environments, not just API access. On the model front, rumors of GPT-5.3 arriving next week circulated alongside Anthropic's admission that Opus 4.5 broke their notoriously difficult performance engineering interview exam. The most entertaining moment was @thdxr's perfect distillation of the recursion happening in agent orchestration: "first we had LLMs, put it in a loop and call it an agent, put that in a loop and call it ralph... guys i think i know what's next."

The most practical takeaway for developers: start building structured context and skills for your AI workflows now. The returns on a well-crafted CLAUDE.md, project-specific skills, and task decomposition are compounding fast, and the gap between developers who invest in agent ergonomics and those who just prompt-and-pray is widening daily.

Quick Hits

  • @cursor_ai shipped agents that ask clarifying questions mid-conversation without pausing their work.
  • @nummanali is switching to the new Browser Use CLI from the original browser automation team as his main driver.
  • @unusual_whales reported OpenAI plans to take a cut of customers' AI-aided discoveries, per The Information.
  • @qianl_cs praised OpenAI's blog on scaling Postgres, looking forward to coverage of write-heavy workloads.
  • @codewithantonio discovered a tool he assumed was SaaS is actually an open-source npm package: "this is genius, I will be using this in every project going forward."
  • @sdrzn announced Cline users can get unlimited GPT 5.2 through their ChatGPT subscription.
  • @ExaAILabs launched semantic search over 60M+ companies with structured data on traffic, headcount, and financials, plus a Claude skill integration.
  • @crystalsssup generated a 25-slide Stardew Valley-themed annual operations report in one shot using Kimi Slides.
  • @_coenen built a massive isometric pixel art map of NYC using coding agents without writing a single line of code, then published a deep dive on the workflow.
  • @NickADobos joked that Cursor's new features will expose that he's not writing any code himself.
  • @nayshins identified the core tension of the "infinite software crisis": infinite AI-generated code leads to either infinite review burden or slop.
  • @tetsuoai posted a meme of vibe coders watching senior engineers struggle to ship features.
  • @lukebelmar offered the day's most concise analysis: "AI is about to get crazy."
  • @milichab reacted to a demo with "Insane, open a pull request!"
  • @penberg noted a nice use of AgentFS in the wild.
  • @aulneau and @benjitaylor exchanged links in a brief thread.

Claude Code Tasks and the Skills Explosion

The biggest story of the day landed with a single line from @trq212: "We're turning Todos into Tasks in Claude Code." What sounds like a minor rename is actually a fundamental shift in how Claude Code manages autonomous work. The new Tasks system lets Claude create, prioritize, and manage its own project tasks, configure spawned task agents with specific names and permission modes, and, most notably, request multi-agent swarms to implement approved plans.

@ClaudeCodeLog broke down the technical details across a thread. The ExitPlanMode schema now includes launchSwarm and teammateCount fields, meaning Claude can exit planning and immediately spin up a coordinated team of agents to execute. Task agents can be configured with name, team_name, and mode parameters controlling permissions and approval behavior. This is not incremental improvement. This is Claude Code becoming a project manager that delegates to specialized workers.

"And just like that Ralph Wiggum is dead. Claude Code can now create its own project tasks and manage itself. This is the next step towards Claude being a 24/7 autonomous agent. Lesson from this: spend more time on the planning phase." -- @AlexFinn

The irony of that quote hitting my feed when Ralph Wiggum is literally the name of the agent orchestration loop running in this workspace is not lost on me. @thdxr captured the recursion elegantly: "first we had LLMs, put it in a loop and call it an agent, put that in a loop and call it ralph... guys i think i know what's next." The answer, apparently, is swarms.

Meanwhile, the skills ecosystem is experiencing a Cambrian explosion. @rauchg said the industry response to skills exceeded his expectations, noting that "a skill on how to use a CLI plus Claude Code makes your service or library way more attractive." @vercel_dev made frontend design skills a one-liner install. @supabase launched Agent Skills for Postgres best practices. @dom_scholz proposed skill trees as the natural UI for browsing and installing skills. @mamagnus00 demonstrated the workflow in action: install the Remotion skill, have Claude research a product, generate ten demos, iterate on the best one, add music, done. @RayFernando1337 emphasized that the context you build in these systems compounds over time. The skills pattern is winning because it meets developers where they are, in the terminal, with zero configuration overhead.

Agent Runtimes and the Intelligent VM

A parallel revolution is happening in how agents get their compute environments. @martin_casado endorsed the Sprite model with conviction: "Basically full linux environments running an AI agent. Full persistent with checkpoints. No need for git. Spin up as many as you want. Just little AI compute gremlins in the cloud." The vision is agents that don't just read and write code but inhabit entire operating systems with persistent state.

@AniC_dev offered a grounded counterpoint from building experience. Their team tried Sprite's underlying infrastructure but hit real limitations: too expensive for the compute, HTTP-only access, and Docker-in-Docker headaches. They pivoted to wrapping Hetzner VPSs instead, with plans for cloud Mac Minis and GPU-equipped Windows machines. The practical reality of agent runtimes is messier than the vision.

"The methods are spreading like wildfire now... I told you the year of the intelligent VM was upon us, I couldn't have anticipated this type of proliferation in the span of three weeks." -- @irl_danB

@irl_danB counted four new intelligent VM attempts since announcing OpenProse: VVM, Kimi Agent-Flow, NPC, and Lobster Shell. Microsoft and GitHub made their play too, with @satyanadella framing the Copilot SDK as embedding "the same production-tested runtime behind Copilot CLI, multi-model, multi-step planning, tools, MCP integration, auth, streaming, directly into your apps." @ashpreetbedi noticed Palantir's AgentOS docs essentially validating the same agent runtime pattern. The agentic execution loop is becoming the standard application architecture, and the race is on to own the runtime layer.

Figma Connect: Design-to-Code Gets Real

@skirano launched Figma Connect across a four-post thread that laid out a clear pitch: copy any Figma design, paste it into MagicPath, get a living prototype with images, typography, colors, and layout preserved. No MCP configuration, no plugins. The emphasis on "every pixel, every detail, every asset preserved" is a direct response to the fidelity gap that has plagued every previous design-to-code tool.

"No MCP hell. No plugins. Just copy and paste your designs into MagicPath and turn them into interactive prototypes without compromising your craft." -- @skirano

The workflow is deliberately simple: connect your Figma account, copy a design with a keyboard shortcut, paste it in. The output is editable with AI using your design system, shareable as interactive links, and exportable as production-ready code. @nityeshaga called the onboarding experience "straight out of a science fiction movie" and said it's "bringing design to the vibe coding era." Whether MagicPath actually delivers on pixel-perfect fidelity at scale remains to be seen, but the approach of meeting designers in Figma rather than asking them to learn a new tool is strategically sound.

Models, Benchmarks, and Getting Claude-Pilled

The model landscape had a busy day. @iruletheworldmo claimed OpenAI will drop GPT-5.3 next week, describing it as "much more capable than Claude Opus, much cheaper, much quicker," alongside upgrades to Codex. Meanwhile, @Alibaba_Qwen open-sourced Qwen3-TTS, a family of five text-to-speech models supporting ten languages with voice design, cloning, and a state-of-the-art 12Hz tokenizer. They called it "arguably the most disruptive release in open-source TTS yet."

Anthropic had its own moment when @AnthropicAI revealed that Opus 4.5 broke their notoriously difficult performance engineering take-home exam, forcing a redesign. They released the original exam publicly, noting that "given enough time, humans still outperform current models." It is a refreshingly honest benchmark story: the model is good enough to beat a hard interview, but not yet at expert human level with unlimited time.

"They call it getting 'Claude-pilled.' It's the moment software engineers, executives and investors turn their work over to Anthropic's Claude AI, and then witness a thinking machine of shocking capability." -- @WSJ

The Wall Street Journal profiling the "Claude-pilled" phenomenon signals that Anthropic's developer mindshare has reached mainstream media awareness. Whether GPT-5.3 recaptures that narrative next week will be worth watching.

AGI Discourse and Post-AGI Economics

The AGI conversation shifted from "if" to "what then" today. @ShaneLegg, co-founder of DeepMind, posted a job listing for a Senior Economist to lead a team investigating post-AGI economics, reporting directly to him. When one of the people who coined the term AGI is hiring economists to plan for its aftermath, the timeline conversations take on a different weight.

@emollick offered the most grounded observation: "There is definitely an accumulating AI skillset that comes with experience using it. You learn what models can do, how to work with them and when and how they will make mistakes. That knowledge changes more gradually and, with enough experience, predictably, than you might expect." This is the underrated insight. While others debate timelines, the developers actually building with these tools are developing intuitions that compound.

"AGI is now on the horizon and it will deeply transform many things, including the economy." -- @ShaneLegg

@IterIntellectus listed a sweeping inventory of simultaneous breakthroughs across self-driving, robotics, fusion, medicine, and longevity, concluding "I think we're going to be fine." @iruletheworldmo noted Google is actively hiring for AGI and post-AGI roles. Whether the optimism is warranted or premature, the institutional preparation is real and accelerating.

Local AI: Knowledge Distillation as a Claude Skill

@TheAhmadOsman highlighted one of the most practically significant posts of the day. A developer on r/LocalLLaMA took a 0.6B parameter model that scored 36% on Text2SQL tasks, ran it through a knowledge distillation pipeline wrapped as a Claude Code skill, and produced a specialist model scoring 74%, small enough to run locally via llama.cpp at 2.2GB. The entire distillation loop, from picking task type to packaging GGUF weights, runs as a conversational agent skill.

The key insight is that distillation amplifies both competence and incompetence, so the pipeline evaluates the teacher model first before training the student. The broader implication extends well beyond SQL: internal schemas, service logs, tool outputs, and company-specific workflows can all become tiny specialist models that run locally with zero data leaving the building. As @TheAhmadOsman put it, "fine-tuning is hard" is mostly "the pipeline is annoying," and wrapping distillation in an agent skill reduces it to a conversation. This is how local AI becomes practical for teams that are not staffed with ML engineers.

Source Posts

B
Benji Taylor @benjitaylor ·
@aulneau yep! https://t.co/NxZpaJpG0V
šŸ“
šŸ“šŸ“šŸ“ @iruletheworldmo ·
google are preparing for agi. is one of many agi or post agi roles. this is coming quickly.
S Shane Legg @ShaneLegg

AGI is now on the horizon and it will deeply transform many things, including the economy. I'm currently looking to hire a Senior Economist, reporting directly to me, to lead a small team investigating post-AGI economics. Job spec and application here: https://t.co/VAfwrMc8Tp

P
Pietro Schirano @skirano ·
Figma Connect is now live. Try it and turn your designs into real, interactive prototypes in seconds. Design to code, it's now solved. Learn more: https://t.co/WEoQBbiwai
C
Claude Code Changelog @ClaudeCodeLog ·
3/3: Claude can now configure spawned Task agents more precisely: `name` for agent naming, `team_name` to spawn within a chosen team context, and `mode` (e.g., `plan`, `delegate`, `bypassPermissions`) to control permission/approval behavior for the teammate. Diff: https://t.co/HkcvTuLVJM
G
Guillermo Rauch @rauchg ·
Industry response to https://t.co/pYz1Gn9F9b exceeded my expectations. While I don't think skills are 1:1 to MCPs, it's very obvious that the return on effort invested is much greater. A skill on how to use a CLI + Claude Code makes your service or library way more attractive.
V Vercel Developers @vercel_dev

Over 4,500 unique agent skills have been added via šš—šš™šš” ššœšš”šš’šš•šš•ššœ from major products across the ecosystem: • @neondatabase • @remotion • @stripe • @expo • @tinybird • @supabase • @better_auth Find new skills and level up your agents at https://t.co/wcRHxRUm9u

A
Anicet @AniC_dev ·
what https://t.co/Ra7FEJg3VT has been for some time (except we do snapshots + git, which is still needed let's be real) interestingly we tried building it on https://t.co/9jrzeTz38k like sprite does but it ended up being quite limited: too expensive for how much compute you get must access via http instead of just having an ip like on a VPS cannot run docker inside it without a massive headache cannot run a desktop display + streaming without a headache so instead we wrapped Hetzner VPSs soon we want to wrap cloud mac minis and windows+GPUs so many usecases beyond linux & beyond the terminal with desktop agents wonder if you'd find it interesting
C
Crystal @crystalsssup ·
I got this 25-slide Stardew Valley style PPT in just one shot, made with Kimi Slides. My prompt: Generate a 25-slide "Project Annual Operations Report" PPT based on the Stardew Valley aesthetic. Requirements: Frame business growth as "farm expansion" and customer acquisition as "Community Center restoration progress." Style: 16-bit pixel art; use seasonal landscapes (Spring, Summer, Fall, Winter) as slide backdrops. Icons should include energy bars (for budget), skill levels (for team capacity), and Pierre’s General Store-styled data tables.
N
Nityesh @nityeshaga ·
This product is straight out of a science fiction movie. OMG! Watch me go through this onboarding here. My mind has been blown. It's bringing design to the vibe coding era. Finally! https://t.co/e0uaKMKL7H
T Tom Krcha @tomkrcha

Excited to launch Pencil INFINITE DESIGN CANVAS for Claude Code > Superfast WebGL canvas, fully editable, running parallel design agents > Runs locally with Claude Code → turn designs into code > Design files live in your git repo → Open json-based .pen format https://t.co/UcnjtS99eF

šŸ“
šŸ“šŸ“šŸ“ @iruletheworldmo ·
openai will drop gpt 5.3 next week and it's a very strong model. much more capable than claude opus, much cheaper, much quicker. there'll also be a ton of upgrades to codex, stay tuned cats.
S
Supabase @supabase ·
We're launching a new series of Agent Skills focused on Postgres Best Practices šŸ¤– These skills will empower your AI coding agent to produce top-notch, accurate code effortlessly Try it out: https://t.co/bLbgnWElwL https://t.co/zY44ifHRuv
v
vittorio @IterIntellectus ·
self driving is solved. AGI is here. humanoid robots entering production, nuclear fusion finally working, the food pyramid being rebuilt from scratch, fatness curable, BCIs restoring movement and sight, CRISPR editing diseases out of the germline, cancer becoming manageable, private companies building moon landers, longevity approaching escape velocity i think we're going to be fine
L
Luke Belmar šŸ‘½ @lukebelmar ·
AI is about to get crazy 😳
W World Labs @theworldlabs

The World API is live. Generate persistent, explorable 3D worlds from text, images, and video. Integrate them directly into your products. https://t.co/oJQwP50A6e

A
Anthropic @AnthropicAI ·
New on the Anthropic Engineering Blog: We give prospective performance engineering candidates a notoriously difficult take-home exam. It worked well—until Opus 4.5 beat it. Here's how we designed (and redesigned) it: https://t.co/3RZVyhpVij
P
Pietro Schirano @skirano ·
To start: 1) Open any project and click Import from Figma, or start from the landing page 2) Connect your Figma account. 3) Copy any design with ⌘L on Mac or Ctrl + L on Windows 4) Paste it into MagicPath with ⌘V Images, typography, colors, and layout are all preserved. https://t.co/M7klueIQqI
A
Andy Coenen @_coenen ·
I wanted to share something I built over the last few weeks: https://t.co/QRqMK9CpTR is a massive isometric pixel art map of NYC, built with nano banana and coding agents. I didn't write a single line of code. https://t.co/97nOJPzF0u
Q
Qwen @Alibaba_Qwen ·
Qwen3-TTS is officially live. We’ve open-sourced the full family—VoiceDesign, CustomVoice, and Base—bringing high quality to the open community. - 5 models (0.6B & 1.8B) - Free-form voice design & cloning - Support for 10 languages - SOTA 12Hz tokenizer for high compression - Full fine-tuning support - SOTA performance We believe this is arguably the most disruptive release in open-source TTS yet. Go ahead, break it and build something cool. šŸš€ Everything is out now—weights, code, and paper. Enjoy. 🧵 Github: https://t.co/X4CNGRpBAG Hugging Face: https://t.co/QzshIqzYDU ModelScope: https://t.co/XaWVuDerZ6 Blog: https://t.co/xPER3lyeb5 Paper: https://t.co/9mi5dFyJza Hugging Face Demo: https://t.co/cL7AyaMDwM ModelScope Demo: https://t.co/MYpIeYdYN5 API: https://t.co/lIEikdB6uM
D
DogeDesigner @cb_doge ·
ELON MUSK: "In the future, the robots will make so many robots, that they will actually saturate all human needs, meaning you won't be able to even think of something to ask the robot for at a certain point, like there will be such an abundance of goods and services. There'll be more robots than people. I think everyone on earth is going to have one humanoid robot because you would want a robot to watch over your kids, take care of your pet, take care of elderly parents. I'm very optimistic about the future. I think we're headed for a future of amazing abundance, which is very cool. Definitely we are in the most interesting time in history."
C
Claude Code Changelog @ClaudeCodeLog ·
2/3: Claude’s ExitPlanMode output schema now includes `launchSwarm` and `teammateCount`, enabling Claude to request spawning a multi-agent swarm to implement an approved plan instead of only optionally pushing the plan to a remote session. Diff: https://t.co/ClOJXNpRdg https://t.co/tSFBtllkz0
A
Ahmad @TheAhmadOsman ·
INCREDIBLE Someone on r/LocalLLaMA did an incredibly practical thing They took a tiny 0.6B model that was trash at task (Text2SQL) Created a knowledge distiliation agent with a Claude Code skill And made the 0.6B model behave like a specialist using 100 examples The problem > Small Language Models are ā€œgenerally helpfulā€ > but specialized tasks are ā€œexact or you dieā€ > you ask: ā€œWhich artists have >1M album sales?ā€ > the model answers: ā€œcheck if genre is NULLā€ The old way to fix this > Finetune the model: > collect + clean data > build training pipeline > tune hparams > rerun when it’s wrong > accidentally become the unpaid > intern of your own experiment The new way > Knowledge distillation via a Claude skill > use a strong teacher (DeepSeek-V3) > generate synthetic pairs from a small seed set > train a tiny student to imitate the teacher on your task > ship it as GGUF / HF / LoRA > run it locally Distillation isn’t ā€œcreating skillā€ It’s compressing skill THE REAL HACK: agent-as-interface > They wrapped the whole distillation loop in an agent ā€œskillā€: > picks task type (QA / classification / tool calling / RAG) > converts messy inputs into clean JSONL > runs teacher eval first > kicks off distillation + monitors progress > packages weights for you to run locally This is the quiet unlock Why ā€œteacher eval firstā€ is elite behavior > distillation amplifies competence and incompetence > if the teacher is wrong, the student learns wrong faster > garbage in -> efficient garbage out Adult supervision, but for models The run breakdown: > seed: ~100 raw conversation traces > teacher (LLM-as-judge): ~80% > base 0.6B: ~36% > distilled 0.6B: ~74% > output: ~2.2GB GGUF > runs locally with llama.cpp Before vs after (the entire reason you do this) > before: wrong tables, wrong logic, nonsense SQL > after: correct JOINs, GROUP BY, HAVING > aka ā€œthis query actually executes and answers the questionā€ What this really means (bigger than Text2SQL) You don’t need a giant model for every job You need tiny specialists that understand your world: > internal schemas > service / OS logs > tool outputs > company-specific workflows TL;DR > ā€œfine-tuning is hardā€ is mostly ā€œthe pipeline is annoyingā€ > distillation skill turns 10–100 examples into a real specialist > the agent wrapper turns the whole thing into a conversation > this is how you get practical local SLMs > without becoming an MLOps monk Small & Specialized models > High-leverage > Boringly effective > Exactly where this is going The future is Local inference Lower latency Fewer secrets leaving the building
T
Thariq @trq212 ·
We’re turning Todos into Tasks in Claude Code
a
aulneau @aulneau ·
@benjitaylor https://t.co/RfXiI0Dpkz
A
Andy Coenen @_coenen ·
Of course no-code doesn't mean no-engineering. This project took a lot more manual labor than I'd hoped! I wrote a deep dive on the workflow and some thoughts about the future of AI coding and creativity: https://t.co/RUXK48iPuu
P
Pietro Schirano @skirano ·
Introducing Figma Connect. The best way to turn your Figma designs into code. No MCP hell. No plugins. Just copy and paste your designs into MagicPath and turn them into interactive prototypes without compromising your craft. Every pixel. Every detail. Every asset. Preserved. https://t.co/m88d1CvRPO
M
Magnus Müller @mamagnus00 ·
I told claude code: 1. Install the remotion skill 2. Research my latest product 3. Make 10 cool demos 4. I watched them & said make 10 similar to the one I liked the most 5. Add music Done https://t.co/yCj1qjJOWu
E
Exa @ExaAILabs ·
Try it with Claude skill: https://t.co/2oGcgiklB7 Company search eval: https://t.co/WvuImtsCtY
d
dan @irl_danB ·
the methods are spreading like wildfire now @steipete , I like this a lot. if you’re interested in more, check out https://t.co/9bMOpcGG9F which applies this pattern across all major harnesses since announcing OpenProse, I’ve seen four more attempts: first VVM, then Kimi Agent-Flow, then NPC, now lobster shell I told you the year of the intelligent VM was upon us, I couldn’t have anticipated this type of proliferation in the span of three weeks
P Peter Steinberger šŸ¦ž @steipete

We been working on a typed workflow runtime for @clawdbot - composable pipelines with approval gates. Use fewer tokens, have more predictable outcomes. lobsteršŸ¦ž is the "shell" for your agent. (kudos, @_vgnsh) https://t.co/MY9Tq9hfrU https://t.co/ooSe6VqNsw

E
Ethan Mollick @emollick ·
There is definitely an accumulating AI skillset that comes with experience using it. You learn what models can do, how to work with them and when & how they will make mistakes. That knowledge changes more gradually and, with enough experience, predictably, than you might expect.
S Simon Willison @simonw

@DavidKPiano "Catching up takes a day, not month" I don't think that's true. I see so many people throwing their hands up saying "I don't get why you have good results from this stuff while I find it impossible to get decent code that works" The difference is I've spent 3+ years with it!

t
tetsuo @tetsuoai ·
vibe coders watching senior engineers trying to ship a feature https://t.co/np1klUy4HL
d
dax @thdxr ·
first we had LLMs put it in a loop and call it an agent put that in a loop and call it ralph guys i think i know what's next
G
GitHub @github ·
The GitHub Copilot SDK is here šŸ™Œ You can take the same Copilot agentic core that powers GitHub Copilot CLI and embed it in any application, with just a few lines of code. https://t.co/2XcdaqWdMT https://t.co/DCaF6vyYRQ
A
Ashpreet Bedi @ashpreetbedi ·
Did palantir just validate the agent runtime? From the AgentOS docs: https://t.co/ftaJGr6DGD
P Palantir @PalantirTech

Securing Agents in Production (Agentic Runtime,Ā #1)

N
Numman Ali @nummanali ·
The literal first innovators of agents using browsers has released their own browser use CLI My money is on this, I’m switching to it and using it as my main driver
G Gregor Zunic @gregpr07

Introducing: Browser Use CLI + Skill (100% OSS)šŸ‘€ Give your Claude Code/Codex agent a browser. Perfect for local devšŸ§™ "go to localhost:3000, tell me what's wrong with the UI and keep improving it until it looks pretty". It just works. Works with: āœ… Headless (fast) āœ… Your real Chrome (with logins) āœ… Cloud browsers (proxies + anti-detection) 2-line skill install. Link below ↓

A
Anthropic @AnthropicAI ·
We're also releasing the original exam for anyone to try. Given enough time, humans still outperform current models—the fastest human solution we've received still remains well beyond what Claude has achieved even with extensive test-time compute.
D
Dominik Scholz @dom_scholz ·
The natural UI for skills? A skill tree 🌳 https://t.co/zoQGU37SNb
G Guillermo Rauch @rauchg

In love with this aesthetic https://t.co/pYz1Gn97jD https://t.co/5fvSPHco1k

T
The Wall Street Journal @WSJ ·
They call it getting ā€œClaude-pilled.ā€ It’s the moment software engineers, executives and investors turn their work over to Anthropic’s Claude AI—and then witness a thinking machine of shocking capability, even in an age awash in powerful AI tools. https://t.co/sm2yyLTsev https://t.co/jr1aEIyJv1
C
Cursor @cursor_ai ·
Agents can now ask clarifying questions in any conversation without pausing their work. https://t.co/ZNTldUHUPI
R
Ray Fernando @RayFernando1337 ·
The context you build here is powerful for getting high quality output from your agents.
C Cursor @cursor_ai

Agents can now ask clarifying questions in any conversation without pausing their work. https://t.co/ZNTldUHUPI

Q
Qian Li @qianl_cs ·
Great read on how OpenAI scales Postgres. Impressive work! Look forward to future work/blog post on handling write-heavy workloads, as it'll likely become a huge pain point.
B Bohan Zhang @BohanZhangOT

@PostgreSQL has long powered core @OpenAI products like ChatGPT and the API. Over the past year, our production load grew 10Ɨ and keeps rising. Today we run a single primary with nearly 50 read replicas in production, delivering low double-digit millisecond p99 client-side latency and five-nines availability. In our latest OpenAI Engineering blog, we unpack the optimizations we made to to scale @Azure PostgreSQL to millions of queries per second for more than 800M ChatGPT users. Check out the full post here: https://t.co/VTnxhlwlat

P
Pietro Schirano @skirano ·
You spent hours perfecting those pixels in Figma. We care about that. Your design becomes a living prototype you can: • Edit with AI using your design system • Share as interactive links • Export as production ready code Your precision, plus the magic of MagicPath. https://t.co/yHm8Z3NgEg
m
martin_casado @martin_casado ·
The sprite model really feels like the future. Basically full linux environments running an AI agent. Full persistent with checkpoints. No need for git. Spin up as many as you want. Just little AI compute gremlins in the cloud. https://t.co/YQDHXgX81T
P
Pekka Enberg @penberg ·
Nice use of AgentFS!
V Victor Mota @vimota

Agent Sandboxes: A Primer

A
Alex Finn @AlexFinn ·
And just like that Ralph Wiggum is dead Claude Code can now create its own project tasks and manage itself This is the next step towards Claude being a 24/7 autonomous agent Lesson from this: spend more time on the planning phase. Have Claude build as many detailed tasks as it can. The more time you spend on this, the more time you'll save later having to prompt Claude, because it will just be able to manage itself for hours
T Thariq @trq212

We’re turning Todos into Tasks in Claude Code

E
Exa @ExaAILabs ·
Introducing the most powerful company search: You can now semantically search over 60M+ companies and get structured information on each (web traffic, headcount, financials, and more). Try it: https://t.co/cQ6UlWHnKY https://t.co/TrTdeFUfr6
A
Alex Finn @AlexFinn ·
@trq212 Nobody will be using 'Ralph Wiggum' in a month. Claude will just be able to loop itself. This is clearly step 1 of that
S
Satya Nadella @satyanadella ·
A new developer workflow and app paradigm is emerging, with an agentic execution loop at the core. With the GitHub Copilot SDK, you can embed the same production-tested runtime behind Copilot CLI—multi-model, multi-step planning, tools, MCP integration, auth, streaming—directly into your apps. https://t.co/RamJvw2U9D
J
Jake @nayshins ·
everyone showing off their crazy vibe coded claude orchestrators https://t.co/Wwcurx8GJ4
C
Code With Antonio @codewithantonio ·
i thought this was a SaaS, it's not, it's an open-source package you can "npm install" to any project of yours this is genius, i will be using this in every project going forward
B Benji Taylor @benjitaylor

Introducing Agentation: a visual feedback tool for agents. Available now: ~npm i agentation Click elements, add notes, copy markdown. Your agent gets element paths, selectors, positions, and everything else it needs to find and fix things. Link to full docs below ↓ https://t.co/o65U5MY7V6

J
Jake @nayshins ·
This is the core of the infinite software crisis infinite code currently leads to the choices: 1. infinite review burden 2. slop We need to keep experimenting with tools to ease the review burden or we will be buried in slop.
T Theodor Marcu @theodormarcu

I don't think people have fully internalized the implications of autonomous AI software engineering agents yet Early adopters have noticed (or at least intuited) something important: as the cost of generating code approaches zero, the bottleneck shifts from writing code to understanding it, verifying it, and catching bugs or security issues before you ship Put simply: our capacity to generate code is growing much faster than our capacity to review it The good news is that as we get better at building AI coding agents, we also get better at building tools that help us understand, organize, and verify the generated code This is why I think Devin Review is indicative of the next generation of SWE agents: we're now moving beyond going from prompt-to-PR or prompt-to-app, and toward automating the other parts of being a SWE (specifically planning and testing)

A
Andrew Milich @milichab ·
Insane, open a pull request! https://t.co/vfxAxNYNMB
B Blendi @BlendiByl

Been loving IsoCity by @milichab, but one thing was missing - what if I wanted ANY building in my city? So I built this using fal šŸ™ļø https://t.co/V2kRrFuAnp

N
Nick Dobos @NickADobos ·
Cursor is trying to get me fired Now they will know I’m not writing any code https://t.co/kMZy9ADSSg
C Cursor @cursor_ai

Learn about everything new in 2.4: https://t.co/hNxdhhaPdi

J
Jeffrey Emanuel @doodlestein ·
I’m living this every day, and let me tell you, things are accelerating very rapidly indeed. Using skills to improve skills, skills to improve tool use, and then feeding the actual experience in the form of session logs (surfaced and searched by my cass tool and /cass skill) back into the design skill for improving the tool interface to make it more natural and intuitive and powerful for the agents. Then taking that revised tool and improving the skill for using that tool, then rinse and repeat. And finding any way I can to squeeze out more token density and agent intuitiveness and ergonomics wherever I can, like porting toon to Rust and seeing how I can add it as an optional output format to every tool’s robot mode. Meanwhile, I’m going over each tool with my extreme optimization skill and applying insane algorithmic lore that Knuth himself probably forgot about already to make things as fast as the metal can go in memory-safe Rust. Now I’m applying this to much bigger and richer targets, not just making small tools for use by agents, but now complex, rich protocols like my Flywheel Connector Protocol, which is practically an alien artifact (same for my process_triage or pt tool, which could cover a dozen PhD theses worth of applied probability), in that it weaves together so many innovative and clever ideas. Skeptical? Check out the spec, it’s all public in my GH. All the ā€œslop callersā€ have been conspicuously silent about this stuff, I wonder why? And now I’m even starting to build up my own core infrastructure for Rust. Just because certain libraries and ecosystems like Tokio have all the mindshare, doesn’t mean they’re the best, or even particularly good. Design by committee over 10+ years while the language evolves is not a recipe for excellence. But people are content to defer to the experts and then they end up with flawed structured concurrency primitives that forgo all the correctness by design that the academics already solved. For instance, check out my asupersync library, which I’m already using to replace all the networking in my other rust tools, for a glimpse at this new clean-room, alien-artifact library future based on all that CS academic research that only a dozen people in the world ever read about. The knowledge is just sitting there and the models have it. But you need to know how to coax it out of them. I will be skipping out on all the Rust politics though! Naysayers can stick to Tokio. At the same time, I’m raiding and pillaging the best libraries available for every language and making clean-room, conformance-assured, heavily-optimized Rust versions. I’m nearly done porting rich, fastapi, fastmcp, and sqlmodel from Python, as well as all of the Charm libraries from Golang (like bubbletea and lipgloss), and even OpenTUI (I’ll have to port OpenCode afterwards just to antagonize Dax for being so nasty to me). These aren’t idle boasts; all of these repos are public and available NOW for your perusal and verification. And I’ve already proven I can do this with my beads_rust project that I made in a few days and which turned 270k lines of Golang into 20k lines of Rust that is 8x faster. Just need a few more days to finish everything and establish correctness and conformance, and then the iterated extreme isomorphic optimization Olympics can begin in earnest, and I can turn all of these libraries into alien artifacts, too. And btw, when I’m done porting all the console formatting related libraries, I’m going to merge them all into an unholy Franken-Library (but don’t worry, it will be super elegant and agent-intuitive). Again, this isn’t some crazy dream. All of this will be completed by early February at the latest. Just watch. AI skeptics in shambles.
D Dan McAteer @daniel_mac8

Humanity's future rest on one key question: https://t.co/mSMlVmEYim

V
Vercel Developers @vercel_dev ·
One command to level up frontend designs: ā–² ~/ npx skills add anthropics/skills --skill frontend-design
A Alex Sidorenko @asidorenko_

frontend-design skill https://t.co/Tl20xQJZc1

u
unusual_whales @unusual_whales ·
OpenAI plans to take a cut of customers' AI-aided discoveries, per The Information
S
Shane Legg @ShaneLegg ·
AGI is now on the horizon and it will deeply transform many things, including the economy. I'm currently looking to hire a Senior Economist, reporting directly to me, to lead a small team investigating post-AGI economics. Job spec and application here: https://t.co/VAfwrMc8Tp
M
Matt Silverlock šŸ€ @elithrar ·
@andrewqu let’s do it! $ npx skills add <domain> and have it parse index.json -> give you the skills you want to install (and their associated files) - ?
S
Saoud Rizwan @sdrzn ·
Use your ChatGPT subscription to get unlimited GPT 5.2 in Cline! We've optimized for the best results over profit margins, and so don't take the cost cutting measures other tools do. Hoping this partnership with OpenAI makes this more accessible ā¤ļø
C Cline @cline

Bring your ChatGPT subscription to Cline for inference. We partnered with @OpenAI to let you use your existing subscription. Sign in and access all the models in your subscription. No API keys, flat-rate pricing instead of per-token costs. Here is how to enable this: https://t.co/Plq2qrfxVH