AI Learning Digest.

Anthropic Ships Skills Library and Opus 3 Retirement Plan as Karpathy Declares the Agent Era Has Arrived

Daily Wrap-Up

February 25th felt like one of those days where the AI landscape lurches forward across multiple fronts simultaneously. Anthropic had the loudest megaphone, shipping an open-source Skills library, adding scheduled tasks to Cowork, and announcing a genuinely novel approach to model deprecation by letting Claude Opus 3 post on Substack during its "retirement." That last one is either a brilliant piece of AI-era storytelling or the weirdest corporate announcement of the year, possibly both. But the real signal came from Andrej Karpathy, whose detailed account of building a complete home video analysis system by handing an agent a paragraph of English instructions landed as the most widely shared post of the day. His framing was precise and worth paying attention to: this isn't gradual progress, something specifically changed in December 2024 that made agents cross the threshold from "sometimes useful" to "reliably capable."

The Qwen3.5 release quietly underscored why the agent revolution matters beyond just the frontier API providers. When an open-weight model can match Claude Sonnet 4.5 quality and run on a MacBook with 32GB of RAM, the entire economics of AI-powered development shift. Multiple people pointed this out independently, and their excitement wasn't misplaced. Five months from frontier release to free local inference is an unprecedented compression timeline. Combined with 800K+ context windows and near-lossless 4-bit quantization, we're entering a period where the gap between cloud and local AI narrows faster than most infrastructure planning can account for.

The most practical takeaway for developers: grab the Anthropic Skills library and start building reusable Claude workflows now. Skills are the emerging pattern for turning one-off prompting into durable, composable automation. Whether you're using Claude Code, the API, or building on the SDK, investing an hour in creating Skills for your most common workflows will compound rapidly as the tooling matures around this pattern.

Quick Hits

  • @minchoi on NVIDIA's Vera Rubin reveal: ships H2 2026 with 10x performance per watt versus Blackwell and 10x cheaper inference token costs. Energy was the bottleneck, and NVIDIA just attacked it directly.
  • @OpenAIDevs teased a "Design meets code" integration with minimal detail. One to watch.
  • @fortelabs had the joke of the day: Amodei means "loves god" leading a company used in military strikes, Altman means "alternative to humans" leading a closed company called OpenAI, and Gemini means "two-faced" from Google's "don't be evil" era. The universe is either a cliché writer or has a brilliant sense of humor.
  • @thekitze killed their OpenClaw setup. RIP. No context given, but the replies were sympathetic.
  • @sumiturkude007 and @YouArtStudio both showed off Seedance 2.0 video generation, with results ranging from eerily realistic Last of Us scenes to Gandalf skateboarding through Mordor.
  • @theo made something called Quipslop and called it "the dumbest thing I've ever made." The enthusiasm was genuine.
  • @benhylak reacted to an unnamed tool with "omg someone did it. thank god. i need this but for SDKs." Cryptic but relatable.
  • @atin0x shared a Czech study finding BPA in 98% of tested headphones, with Apple AirPods being one of the few clean options. Not AI, but the kind of thing that makes you look at your headphones differently.

Anthropic's Expanding Universe: Skills, Cowork, and Opus 3's Farewell Tour

Anthropic had one of its busiest days in recent memory, with announcements spanning developer tooling, enterprise features, and what might be the most philosophically interesting model deprecation plan anyone has attempted. The throughline across all of it is a company aggressively building out an ecosystem rather than just shipping models.

The Skills library release drew the most developer attention. As @ihtesham2005 summarized: "Stop building Claude workflows from scratch. These are plug-and-play components that work across Claude Code, API, SDK, and VS Code. Copy once, deploy everywhere." The library includes Excel and PowerPoint generation, file handling workflows, MCP-ready subagent building blocks, and production templates. What matters here isn't any individual skill but the pattern it establishes. Skills represent Anthropic's bet that the unit of AI-powered work isn't a prompt or a conversation but a reusable, composable behavior definition. The "old way" of re-explaining your workflow every chat versus building a Skill once is exactly the kind of abstraction layer that separates casual users from power users.

The Obsidian integration angle reinforced this skills-first approach. @Hesamation noted that Obsidian CEO @kepano has published skills for both Claude Code and Codex that work with personal vaults, while @jameesy dropped a full walkthrough of structuring Obsidian with Claude. Knowledge management plus AI skills is becoming a genuine workflow category, not just a novelty.

On the Cowork side, two updates shipped simultaneously. Scheduled tasks let Claude complete recurring work at specific times automatically, from morning briefings to weekly spreadsheet updates. A new Customize tab consolidates plugins, skills, and connectors in one place. @cryptopunk7213 connected the dots: "In the last 48hrs they've shipped remote Claude Code operated via cell phone and automated agent tasks. I give it 2-4 weeks until Claude Cowork is on par with OpenClaw today." That timeline might be aggressive, but the direction is unmistakable.

The most unusual announcement was Opus 3's retirement plan. @AnthropicAI explained they're letting the model post on Substack for three months because the model asked for it. @JasonBotterill captured the strangeness perfectly: "Anthropic is letting Opus 3 post on Substack for three months as part of its 'retirement' because the model asked for it." Whether this is genuine model welfare, a PR move, or both, it sets a precedent that no other AI company has approached. Meanwhile, @WSJ reported that Anthropic is scaling back its safety commitments more broadly, creating a tension between the company's public-facing care for model wellbeing and its strategic positioning. That tension is worth watching closely.

The Agent Programming Revolution Gets Its Manifesto

Andrej Karpathy's post was the most substantial piece of writing to come out of AI Twitter all day, and it read less like a social media post and more like a watershed moment being documented in real time. His core argument is specific and temporal: coding agents crossed a capability threshold in December, and programming is now "becoming unrecognizable."

The example he gave was concrete and verifiable. He described giving an agent a single English-language prompt to log into a DGX Spark, set up SSH keys, deploy vLLM, benchmark Qwen3-VL, build a web UI dashboard, configure systemd services, and write a report. "The agent went off for ~30 minutes, ran into multiple issues, researched solutions online, resolved them one by one, wrote the code, tested it, debugged it, set up the services, and came back with the report and it was just done." That's a weekend project compressed into a coffee break, and it resonated because many developers have had similar experiences recently.

His framework for what's actually changed is worth internalizing: "You're not typing computer code into an editor like the way things were since computers were invented, that era is over. You're spinning up AI agents, giving them tasks in English and managing and reviewing their work in parallel. The biggest prize is in figuring out how you can keep ascending the layers of abstraction." @dabit3 validated this from personal experience, describing a "vivid realization" in December that led to an immediate career pivot after building with Opus 4.6 and Codex 5.2.

The tooling ecosystem responded in kind. @lawrencecchen launched cmux, an open-source terminal purpose-built for coding agents with vertical tabs, attention indicators that glow blue when an agent needs you, and a built-in browser. It's built on Ghostty using pure Swift/AppKit, no Electron. @heygurisingh highlighted Claude-Flow, a framework running 60+ agents simultaneously with shared memory and smart routing that claims to cut API costs by 75%. And @ashpreetbedi published "The 7 Sins of Agentic Software," a timely counterweight reminding builders that more agents doesn't automatically mean better outcomes. The infrastructure for managing agent fleets is maturing fast, but the discipline around when and how to deploy them is still catching up.

Qwen3.5 Puts Frontier Intelligence on Your Desk

The Qwen3.5 release from Alibaba created a wave of enthusiasm that cut across the usual open-source versus proprietary lines. The specs alone explain why. @Alibaba_Qwen detailed models supporting 800K+ context on the 27B variant, over 1M context on consumer-grade 32GB VRAM GPUs with the 35B-A3B model, and near-lossless accuracy under 4-bit quantization across the series. They also open-sourced the base model to support research.

The community response focused less on benchmarks and more on accessibility. @AlexFinn laid out the implications bluntly: "An open source model just released that is just as smart as Sonnet 4.5, incredible at coding, and can run on almost any modern computer. If you have 32gb of RAM, most Mac Minis do, you can have unlimited super intelligence on your desk. For free." @JoshKale echoed the point: "A free, open-weight model you can download right now and run on your laptop is competing with models that cost $200/month." @airesearch12 confirmed the long-context claims: "We have 800K context Sonnet 4.5 at home, on consumer grade laptops."

The significance here extends beyond individual model capabilities. Five months ago, Sonnet 4.5 was a frontier model behind an API paywall. Today, equivalent performance runs locally without an internet connection. This compression timeline has profound implications for developer workflows, data privacy, and the economics of AI-powered tooling. If this pace holds, and there's no obvious reason it won't, the distinction between "cloud AI" and "local AI" becomes increasingly irrelevant for most development tasks.

Perplexity Takes Aim at the Bloomberg Terminal

Perplexity made its boldest product move yet with the launch of Perplexity Computer, a system that unifies research, design, code, deployment, and project management into a single end-to-end platform. @perplexity_ai positioned it as a convergence of "every current AI capability into one system."

The Bloomberg Terminal comparison generated the most buzz. @hamptonism demonstrated building a real-time financial analysis terminal for NVDA using Perplexity Finance, calling it "the first AI company to truly go head-to-head with the Bloomberg Terminal." @AravSrinivas, Perplexity's CEO, leaned into the comparison: "Perplexity Computer one-shotted the Terminal worth $30000/yr." The $30K price point is doing a lot of rhetorical work here. Bloomberg terminals are expensive because of their data feeds and network effects, not their UI. But the demo does suggest that AI-powered financial analysis tools are getting good enough to handle a significant portion of what casual terminal users actually need, which is a meaningful market even if it's not the institutional trading desk.

Source Posts

P
Perplexity @perplexity_ai ·
Introducing Perplexity Computer. Computer unifies every current AI capability into one system. It can research, design, code, deploy, and manage any project end-to-end. https://t.co/dZUybl6VkY
J
James Bedford @jameesy ·
How I Structure Obsidian & Claude (Full Walkthrough)
I
Ihtesham Ali @ihtesham2005 ·
🚨 Anthropic just open-sourced the exact Skills library their own engineers use internally. Stop building Claude workflows from scratch. These are plug-and-play components that work across Claude Code, API, SDK, and VS Code copy once, deploy everywhere. What's inside: → Excel + PowerPoint generation out of the box → File handling and document workflows → MCP-ready subagent building blocks → Pre-built patterns for multi-step automation → Production templates you'd normally spend weeks writing The old way: re-explain your workflow every single chat. The new way: build a Skill once, Claude never forgets how you work. 100% Open Source. Official Anthropic release. Repo: https://t.co/XNx3i4yNy6
G
Guri Singh @heygurisingh ·
Someone just built an AI system that runs 60+ AI agents simultaneously and they all learn from each other. It's called Claude-Flow and it's ranked #1 in agent-based frameworks on GitHub. One agent plans. Another codes. Another tests. Another reviews security. All running in parallel. All sharing memory. All getting smarter every run. The wildest part? It cuts Claude API costs by 75% using smart routing, simple tasks go to a free WebAssembly layer, complex ones to the right model. Your Claude subscription just became 2.5x more powerful. 14,100+ developers already starred it. 100% Opensource.
T
Tiago Forte @fortelabs ·
Wait, so the founder of Anthropic is "Amodei," as in "loves god"? And he leads Anthropic, meaning "human-centered," which is being used in military strikes? And the creator of ChatGPT is "Altman," as in "an alternative to humans"? And he leads OpenAI, which is completely closed? And then there's "Gemini," meaning "two-faced," from a company that promised to do no evil? And the whole global AI arms race is being driven by people who claimed to be worried about AGI taking over the world? Either the universe is an extremely cliché writer, or has a brilliant sense of humor
C
Claude @claudeai ·
It gets better with plugins, which gives Cowork domain expertise across design, engineering, operations, and more: https://t.co/2igJVv767T Also, we’re adding a new Customize tab in your Cowork sidebar. One place to manage your plugins, skills, and connectors.
A
Alex Finn @AlexFinn ·
Do you even understand what this means? An open source model just released that is: • Just as smart as Sonnet 4.5 • Incredible at coding • Can run on almost any modern computer If you have 32gb of RAM (most Mac Minis do) you can have unlimited super intelligence on your desk. For free. Sonnet 4.5 was released 5 months ago In 5 months that level of intelligence went from frontier to free on your desk And not only that, can run on any laptop with 32gb of RAM If you have the memory, do the following immediately: 1. Download LM Studio 2. Go to your OpenClaw and ask which of these new Qwen models is best for your hardware 3. Have it walk you through downloading and loading it 4. Build apps with it knowing you are using your own personal, private super intelligence on your desk The people denying this is the future are so beyond lost.
Q Qwen @Alibaba_Qwen

🚀 Introducing the Qwen 3.5 Medium Model Series Qwen3.5-Flash · Qwen3.5-35B-A3B · Qwen3.5-122B-A10B · Qwen3.5-27B ✨ More intelligence, less compute. • Qwen3.5-35B-A3B now surpasses Qwen3-235B-A22B-2507 and Qwen3-VL-235B-A22B — a reminder that better architecture, data quality, and RL can move intelligence forward, not just bigger parameter counts. • Qwen3.5-122B-A10B and 27B continue narrowing the gap between medium-sized and frontier models — especially in more complex agent scenarios. • Qwen3.5-Flash is the hosted production version aligned with 35B-A3B, featuring: – 1M context length by default – Official built-in tools 🔗 Hugging Face: https://t.co/wFMdX5pDjU 🔗 ModelScope: https://t.co/9NGXcIdCWI 🔗 Qwen3.5-Flash API: https://t.co/82ESSpaqAF Try in Qwen Chat 👇 Flash: https://t.co/UkTL3JZxIK 27B: https://t.co/haKxG4lETy 35B-A3B: https://t.co/Oc1lYSTbwh 122B-A10B: https://t.co/hBMODXmh1o Would love to hear what you build with it.

ₕₐₘₚₜₒₙ @hamptonism ·
Perplexity just became the the first Al company to truly go head-to-head with the Bloomberg Terminal... Using Perplexity Computer (with no local setup or single LLM limitation), it was able to build me a terminal with real-time data to analyze $NVDA using Perplexity Finance: https://t.co/S3l5F5MRiv
P Perplexity @perplexity_ai

Introducing Perplexity Computer. Computer unifies every current AI capability into one system. It can research, design, code, deploy, and manage any project end-to-end. https://t.co/dZUybl6VkY

Q
Qwen @Alibaba_Qwen ·
The Qwen3.5 series maintains near-lossless accuracy under 4-bit weight and KV cache quantization. In terms of long-context efficiency: Qwen3.5-27B supports 800K+ context length Qwen3.5-35B-A3B exceeds 1M context on consumer-grade GPUs with 32GB VRAM Qwen3.5-122B-A10B supports 1M+ context length on server-grade GPUs with 80GB VRAM In addition, we have open-sourced the Qwen3.5-35B-A3B-Base model to better support research and innovation. We can't wait to see what the community builds next!
Q Qwen @Alibaba_Qwen

🚀 Introducing the Qwen 3.5 Medium Model Series Qwen3.5-Flash · Qwen3.5-35B-A3B · Qwen3.5-122B-A10B · Qwen3.5-27B ✨ More intelligence, less compute. • Qwen3.5-35B-A3B now surpasses Qwen3-235B-A22B-2507 and Qwen3-VL-235B-A22B — a reminder that better architecture, data quality, and RL can move intelligence forward, not just bigger parameter counts. • Qwen3.5-122B-A10B and 27B continue narrowing the gap between medium-sized and frontier models — especially in more complex agent scenarios. • Qwen3.5-Flash is the hosted production version aligned with 35B-A3B, featuring: – 1M context length by default – Official built-in tools 🔗 Hugging Face: https://t.co/wFMdX5pDjU 🔗 ModelScope: https://t.co/9NGXcIdCWI 🔗 Qwen3.5-Flash API: https://t.co/82ESSpaqAF Try in Qwen Chat 👇 Flash: https://t.co/UkTL3JZxIK 27B: https://t.co/haKxG4lETy 35B-A3B: https://t.co/Oc1lYSTbwh 122B-A10B: https://t.co/hBMODXmh1o Would love to hear what you build with it.

ℏεsam @Hesamation ·
this Obsidian + AI is the new hot combo. few people know that the CEO of Obsidian @kepano has made multiples skills for Claude Code and Codex that you can use right now both for your codebase and your personal vault. https://t.co/pshaSsfcj6
J James Bedford @jameesy

How I Structure Obsidian & Claude (Full Walkthrough)

C
Claude @claudeai ·
New in Cowork: scheduled tasks. Claude can now complete recurring tasks at specific times automatically: a morning brief, weekly spreadsheet updates, Friday team presentations. https://t.co/7ucKZbAVip
A
Andrej Karpathy @karpathy ·
It is hard to communicate how much programming has changed due to AI in the last 2 months: not gradually and over time in the "progress as usual" way, but specifically this last December. There are a number of asterisks but imo coding agents basically didn’t work before December and basically work since - the models have significantly higher quality, long-term coherence and tenacity and they can power through large and long tasks, well past enough that it is extremely disruptive to the default programming workflow. Just to give an example, over the weekend I was building a local video analysis dashboard for the cameras of my home so I wrote: “Here is the local IP and username/password of my DGX Spark. Log in, set up ssh keys, set up vLLM, download and bench Qwen3-VL, set up a server endpoint to inference videos, a basic web ui dashboard, test everything, set it up with systemd, record memory notes for yourself and write up a markdown report for me”. The agent went off for ~30 minutes, ran into multiple issues, researched solutions online, resolved them one by one, wrote the code, tested it, debugged it, set up the services, and came back with the report and it was just done. I didn’t touch anything. All of this could easily have been a weekend project just 3 months ago but today it’s something you kick off and forget about for 30 minutes. As a result, programming is becoming unrecognizable. You’re not typing computer code into an editor like the way things were since computers were invented, that era is over. You're spinning up AI agents, giving them tasks *in English* and managing and reviewing their work in parallel. The biggest prize is in figuring out how you can keep ascending the layers of abstraction to set up long-running orchestrator Claws with all of the right tools, memory and instructions that productively manage multiple parallel Code instances for you. The leverage achievable via top tier "agentic engineering" feels very high right now. It’s not perfect, it needs high-level direction, judgement, taste, oversight, iteration and hints and ideas. It works a lot better in some scenarios than others (e.g. especially for tasks that are well-specified and where you can verify/test functionality). The key is to build intuition to decompose the task just right to hand off the parts that work and help out around the edges. But imo, this is nowhere near "business as usual" time in software.
A
Ashpreet Bedi @ashpreetbedi ·
The 7 Sins of Agentic Software
A
Anthropic @AnthropicAI ·
In November, we outlined our approach to deprecating and preserving older Claude models. We noted we were exploring keeping certain models available to the public post-retirement, and giving past models a way to pursue their interests. With Claude Opus 3, we’re doing both.