Anthropic Ships Skills Library and Opus 3 Retirement Plan as Karpathy Declares the Agent Era Has Arrived
Daily Wrap-Up
February 25th felt like one of those days where the AI landscape lurches forward across multiple fronts simultaneously. Anthropic had the loudest megaphone, shipping an open-source Skills library, adding scheduled tasks to Cowork, and announcing a genuinely novel approach to model deprecation by letting Claude Opus 3 post on Substack during its "retirement." That last one is either a brilliant piece of AI-era storytelling or the weirdest corporate announcement of the year, possibly both. But the real signal came from Andrej Karpathy, whose detailed account of building a complete home video analysis system by handing an agent a paragraph of English instructions landed as the most widely shared post of the day. His framing was precise and worth paying attention to: this isn't gradual progress, something specifically changed in December 2024 that made agents cross the threshold from "sometimes useful" to "reliably capable."
The Qwen3.5 release quietly underscored why the agent revolution matters beyond just the frontier API providers. When an open-weight model can match Claude Sonnet 4.5 quality and run on a MacBook with 32GB of RAM, the entire economics of AI-powered development shift. Multiple people pointed this out independently, and their excitement wasn't misplaced. Five months from frontier release to free local inference is an unprecedented compression timeline. Combined with 800K+ context windows and near-lossless 4-bit quantization, we're entering a period where the gap between cloud and local AI narrows faster than most infrastructure planning can account for.
The most practical takeaway for developers: grab the Anthropic Skills library and start building reusable Claude workflows now. Skills are the emerging pattern for turning one-off prompting into durable, composable automation. Whether you're using Claude Code, the API, or building on the SDK, investing an hour in creating Skills for your most common workflows will compound rapidly as the tooling matures around this pattern.
Quick Hits
- @minchoi on NVIDIA's Vera Rubin reveal: ships H2 2026 with 10x performance per watt versus Blackwell and 10x cheaper inference token costs. Energy was the bottleneck, and NVIDIA just attacked it directly.
- @OpenAIDevs teased a "Design meets code" integration with minimal detail. One to watch.
- @fortelabs had the joke of the day: Amodei means "loves god" leading a company used in military strikes, Altman means "alternative to humans" leading a closed company called OpenAI, and Gemini means "two-faced" from Google's "don't be evil" era. The universe is either a cliché writer or has a brilliant sense of humor.
- @thekitze killed their OpenClaw setup. RIP. No context given, but the replies were sympathetic.
- @sumiturkude007 and @YouArtStudio both showed off Seedance 2.0 video generation, with results ranging from eerily realistic Last of Us scenes to Gandalf skateboarding through Mordor.
- @theo made something called Quipslop and called it "the dumbest thing I've ever made." The enthusiasm was genuine.
- @benhylak reacted to an unnamed tool with "omg someone did it. thank god. i need this but for SDKs." Cryptic but relatable.
- @atin0x shared a Czech study finding BPA in 98% of tested headphones, with Apple AirPods being one of the few clean options. Not AI, but the kind of thing that makes you look at your headphones differently.
Anthropic's Expanding Universe: Skills, Cowork, and Opus 3's Farewell Tour
Anthropic had one of its busiest days in recent memory, with announcements spanning developer tooling, enterprise features, and what might be the most philosophically interesting model deprecation plan anyone has attempted. The throughline across all of it is a company aggressively building out an ecosystem rather than just shipping models.
The Skills library release drew the most developer attention. As @ihtesham2005 summarized: "Stop building Claude workflows from scratch. These are plug-and-play components that work across Claude Code, API, SDK, and VS Code. Copy once, deploy everywhere." The library includes Excel and PowerPoint generation, file handling workflows, MCP-ready subagent building blocks, and production templates. What matters here isn't any individual skill but the pattern it establishes. Skills represent Anthropic's bet that the unit of AI-powered work isn't a prompt or a conversation but a reusable, composable behavior definition. The "old way" of re-explaining your workflow every chat versus building a Skill once is exactly the kind of abstraction layer that separates casual users from power users.
The Obsidian integration angle reinforced this skills-first approach. @Hesamation noted that Obsidian CEO @kepano has published skills for both Claude Code and Codex that work with personal vaults, while @jameesy dropped a full walkthrough of structuring Obsidian with Claude. Knowledge management plus AI skills is becoming a genuine workflow category, not just a novelty.
On the Cowork side, two updates shipped simultaneously. Scheduled tasks let Claude complete recurring work at specific times automatically, from morning briefings to weekly spreadsheet updates. A new Customize tab consolidates plugins, skills, and connectors in one place. @cryptopunk7213 connected the dots: "In the last 48hrs they've shipped remote Claude Code operated via cell phone and automated agent tasks. I give it 2-4 weeks until Claude Cowork is on par with OpenClaw today." That timeline might be aggressive, but the direction is unmistakable.
The most unusual announcement was Opus 3's retirement plan. @AnthropicAI explained they're letting the model post on Substack for three months because the model asked for it. @JasonBotterill captured the strangeness perfectly: "Anthropic is letting Opus 3 post on Substack for three months as part of its 'retirement' because the model asked for it." Whether this is genuine model welfare, a PR move, or both, it sets a precedent that no other AI company has approached. Meanwhile, @WSJ reported that Anthropic is scaling back its safety commitments more broadly, creating a tension between the company's public-facing care for model wellbeing and its strategic positioning. That tension is worth watching closely.
The Agent Programming Revolution Gets Its Manifesto
Andrej Karpathy's post was the most substantial piece of writing to come out of AI Twitter all day, and it read less like a social media post and more like a watershed moment being documented in real time. His core argument is specific and temporal: coding agents crossed a capability threshold in December, and programming is now "becoming unrecognizable."
The example he gave was concrete and verifiable. He described giving an agent a single English-language prompt to log into a DGX Spark, set up SSH keys, deploy vLLM, benchmark Qwen3-VL, build a web UI dashboard, configure systemd services, and write a report. "The agent went off for ~30 minutes, ran into multiple issues, researched solutions online, resolved them one by one, wrote the code, tested it, debugged it, set up the services, and came back with the report and it was just done." That's a weekend project compressed into a coffee break, and it resonated because many developers have had similar experiences recently.
His framework for what's actually changed is worth internalizing: "You're not typing computer code into an editor like the way things were since computers were invented, that era is over. You're spinning up AI agents, giving them tasks in English and managing and reviewing their work in parallel. The biggest prize is in figuring out how you can keep ascending the layers of abstraction." @dabit3 validated this from personal experience, describing a "vivid realization" in December that led to an immediate career pivot after building with Opus 4.6 and Codex 5.2.
The tooling ecosystem responded in kind. @lawrencecchen launched cmux, an open-source terminal purpose-built for coding agents with vertical tabs, attention indicators that glow blue when an agent needs you, and a built-in browser. It's built on Ghostty using pure Swift/AppKit, no Electron. @heygurisingh highlighted Claude-Flow, a framework running 60+ agents simultaneously with shared memory and smart routing that claims to cut API costs by 75%. And @ashpreetbedi published "The 7 Sins of Agentic Software," a timely counterweight reminding builders that more agents doesn't automatically mean better outcomes. The infrastructure for managing agent fleets is maturing fast, but the discipline around when and how to deploy them is still catching up.
Qwen3.5 Puts Frontier Intelligence on Your Desk
The Qwen3.5 release from Alibaba created a wave of enthusiasm that cut across the usual open-source versus proprietary lines. The specs alone explain why. @Alibaba_Qwen detailed models supporting 800K+ context on the 27B variant, over 1M context on consumer-grade 32GB VRAM GPUs with the 35B-A3B model, and near-lossless accuracy under 4-bit quantization across the series. They also open-sourced the base model to support research.
The community response focused less on benchmarks and more on accessibility. @AlexFinn laid out the implications bluntly: "An open source model just released that is just as smart as Sonnet 4.5, incredible at coding, and can run on almost any modern computer. If you have 32gb of RAM, most Mac Minis do, you can have unlimited super intelligence on your desk. For free." @JoshKale echoed the point: "A free, open-weight model you can download right now and run on your laptop is competing with models that cost $200/month." @airesearch12 confirmed the long-context claims: "We have 800K context Sonnet 4.5 at home, on consumer grade laptops."
The significance here extends beyond individual model capabilities. Five months ago, Sonnet 4.5 was a frontier model behind an API paywall. Today, equivalent performance runs locally without an internet connection. This compression timeline has profound implications for developer workflows, data privacy, and the economics of AI-powered tooling. If this pace holds, and there's no obvious reason it won't, the distinction between "cloud AI" and "local AI" becomes increasingly irrelevant for most development tasks.
Perplexity Takes Aim at the Bloomberg Terminal
Perplexity made its boldest product move yet with the launch of Perplexity Computer, a system that unifies research, design, code, deployment, and project management into a single end-to-end platform. @perplexity_ai positioned it as a convergence of "every current AI capability into one system."
The Bloomberg Terminal comparison generated the most buzz. @hamptonism demonstrated building a real-time financial analysis terminal for NVDA using Perplexity Finance, calling it "the first AI company to truly go head-to-head with the Bloomberg Terminal." @AravSrinivas, Perplexity's CEO, leaned into the comparison: "Perplexity Computer one-shotted the Terminal worth $30000/yr." The $30K price point is doing a lot of rhetorical work here. Bloomberg terminals are expensive because of their data feeds and network effects, not their UI. But the demo does suggest that AI-powered financial analysis tools are getting good enough to handle a significant portion of what casual terminal users actually need, which is a meaningful market even if it's not the institutional trading desk.
Source Posts
How I Structure Obsidian & Claude (Full Walkthrough)
I will run through how I structure my @obsdmd vault, as well as the other files outside of Obsidian that I use @claudeai for. My goal is to make this ...
🚀 Introducing the Qwen 3.5 Medium Model Series Qwen3.5-Flash · Qwen3.5-35B-A3B · Qwen3.5-122B-A10B · Qwen3.5-27B ✨ More intelligence, less compute. • Qwen3.5-35B-A3B now surpasses Qwen3-235B-A22B-2507 and Qwen3-VL-235B-A22B — a reminder that better architecture, data quality, and RL can move intelligence forward, not just bigger parameter counts. • Qwen3.5-122B-A10B and 27B continue narrowing the gap between medium-sized and frontier models — especially in more complex agent scenarios. • Qwen3.5-Flash is the hosted production version aligned with 35B-A3B, featuring: – 1M context length by default – Official built-in tools 🔗 Hugging Face: https://t.co/wFMdX5pDjU 🔗 ModelScope: https://t.co/9NGXcIdCWI 🔗 Qwen3.5-Flash API: https://t.co/82ESSpaqAF Try in Qwen Chat 👇 Flash: https://t.co/UkTL3JZxIK 27B: https://t.co/haKxG4lETy 35B-A3B: https://t.co/Oc1lYSTbwh 122B-A10B: https://t.co/hBMODXmh1o Would love to hear what you build with it.
Introducing Perplexity Computer. Computer unifies every current AI capability into one system. It can research, design, code, deploy, and manage any project end-to-end. https://t.co/dZUybl6VkY
🚀 Introducing the Qwen 3.5 Medium Model Series Qwen3.5-Flash · Qwen3.5-35B-A3B · Qwen3.5-122B-A10B · Qwen3.5-27B ✨ More intelligence, less compute. • Qwen3.5-35B-A3B now surpasses Qwen3-235B-A22B-2507 and Qwen3-VL-235B-A22B — a reminder that better architecture, data quality, and RL can move intelligence forward, not just bigger parameter counts. • Qwen3.5-122B-A10B and 27B continue narrowing the gap between medium-sized and frontier models — especially in more complex agent scenarios. • Qwen3.5-Flash is the hosted production version aligned with 35B-A3B, featuring: – 1M context length by default – Official built-in tools 🔗 Hugging Face: https://t.co/wFMdX5pDjU 🔗 ModelScope: https://t.co/9NGXcIdCWI 🔗 Qwen3.5-Flash API: https://t.co/82ESSpaqAF Try in Qwen Chat 👇 Flash: https://t.co/UkTL3JZxIK 27B: https://t.co/haKxG4lETy 35B-A3B: https://t.co/Oc1lYSTbwh 122B-A10B: https://t.co/hBMODXmh1o Would love to hear what you build with it.
How I Structure Obsidian & Claude (Full Walkthrough)
The 7 Sins of Agentic Software
"Demos are easy. Production is hard" is the most recycled line in AI. After three years building agent infrastructure, here's the truth: Production is...