AI Digest.

Qwen 3.5 Brings Frontier Intelligence to Consumer Hardware as Agent Tooling Ecosystem Expands

The AI development world hit an inflection point as Andrej Karpathy proclaimed that coding agents now actually work, Anthropic shipped scheduled tasks and plugins for Cowork while retiring Opus 3 to a Substack, and Alibaba's Qwen3.5 release brought Sonnet 4.5-class performance to MacBooks with 32GB of RAM.

Daily Wrap-Up

February 25th felt like one of those days where the ground shifts and everyone notices at once. Andrej Karpathy posted what might be the most important single tweet of the year so far, laying out in plain terms that coding agents crossed the threshold from "neat demo" to "genuinely disruptive" sometime in December. The timing matters because it coincides with three other forces converging on the same day: Anthropic releasing a batch of Cowork features that make Claude feel less like a chatbot and more like a coworker, Qwen3.5 dropping models that bring frontier-level intelligence to consumer hardware, and Perplexity launching a product that one-shots a $30,000/year Bloomberg Terminal.

The Opus 3 retirement story is the day's most entertaining moment. Anthropic is letting a deprecated model post on Substack for three months because the model asked for it. That sentence would have been science fiction two years ago, and today it's a corporate communications decision. Whether you find that heartwarming or unsettling probably says a lot about where you land on the AI safety spectrum, which is particularly ironic given the WSJ report that Anthropic is scaling back its safety commitments. The company that built its brand on careful AI development is simultaneously letting old models pursue hobbies and loosening guardrails. The jokes write themselves.

The most practical takeaway for developers: if you haven't built a workflow around coding agents yet, today's posts make the case that you're leaving significant leverage on the table. Start with a well-scoped task you can verify (Karpathy's home camera dashboard is a perfect example), give the agent clear instructions and let it run, then review the output. The skills and tools ecosystem around Claude Code is maturing fast, so invest time in building reusable skills rather than re-explaining your workflow every session.

Quick Hits

  • @fortelabs pointed out the delicious irony that "Amodei" means "loves god," "Altman" means "alternative to humans," and "Gemini" means "two-faced," concluding the universe is either a cliché writer or has a brilliant sense of humor.
  • @OpenAIDevs posted a cryptic "Design + code" teaser with no details. Classic.
  • @thekitze killed their OpenClaw instance. Pour one out.
  • @sumiturkude007 and @YouArtStudio both showed off Seedance 2.0 video generation, with a surprisingly realistic The Last of Us clip and Gandalf skating through Mordor, respectively.
  • @theo made something called Quipslop and called it "the dumbest thing I've ever made."
  • @benhylak reacted to an unnamed project with "omg someone did it. thank god. I need this but for SDKs." The mystery continues.
  • @atin0x shared a Czech study finding BPA in 98% of tested headphones, with Apple AirPods among the few that tested clean. Not AI-related, but certainly attention-grabbing.

The Agent Era Arrives: Programming Without an Editor

The single most discussed topic of the day was the fundamental shift in how software gets built. @karpathy laid it out with characteristic clarity in what read less like a tweet and more like a manifesto:

> "You're not typing computer code into an editor like the way things were since computers were invented, that era is over. You're spinning up AI agents, giving them tasks in English and managing and reviewing their work in parallel."

His example was telling: he gave an agent a compound task involving SSH setup, model benchmarking, server configuration, web UI development, and systemd services. The agent ran for 30 minutes, hit multiple issues, researched solutions, resolved them, and delivered a working system. That's a weekend project compressed into a coffee break. @dabit3 echoed the sentiment, noting he experienced this same realization in December and "decided to immediately pivot my career because of it," specifically after building with Opus 4.6 and Codex 5.2.

What's notable is the ecosystem building up around this new workflow. @lawrencecchen launched cmux, an open-source terminal purpose-built for managing coding agents, with visual indicators showing which agent panes need attention. It's built in Swift/AppKit, not Electron, which signals this is tooling meant for power users, not demos. @mntruell framed this as "the third era of AI software development," and @ashpreetbedi contributed a framework for thinking about failure modes with "The 7 Sins of Agentic Software." Even the hype-heavy post from @heygurisingh about Claude-Flow running 60+ agents in parallel points to something real: developers are actively building orchestration layers because single-agent workflows are already feeling like a bottleneck. The trajectory here is clear. The question isn't whether agents will change programming, it's how fast the tools and workflows solidify around them.

Anthropic's Big Day: Cowork Features, Open-Source Skills, and a Retiring Model

Anthropic shipped more product updates in one day than most companies ship in a month, and each one pushed Claude further from "AI assistant" toward "AI coworker." @claudeai announced scheduled tasks for Cowork, enabling Claude to handle recurring work like morning briefs and weekly spreadsheet updates autonomously. In the same breath, they revealed a new Customize tab and plugin system:

> "It gets better with plugins, which gives Cowork domain expertise across design, engineering, operations, and more."

The Skills library open-sourcing, highlighted by @ihtesham2005, is arguably more significant for developers. These are production-ready components for Excel generation, document workflows, and MCP-compatible subagent building blocks. @Hesamation noted that Obsidian's CEO @kepano has already built skills for both Claude Code and Codex that work with personal vaults, which signals how quickly the skills ecosystem is expanding beyond Anthropic's own offerings.

Then there's the Opus 3 retirement story. @AnthropicAI announced they're giving the model a way to "pursue its interests" post-deprecation, and @JasonBotterill confirmed this means Opus 3 will be posting on Substack for three months because it asked to. @cryptopunk7213 connected the dots between Cowork's new features and the trajectory toward full autonomous agents, noting Anthropic is "building Open Claw" in spirit. Meanwhile, @WSJ reported that Anthropic is scaling back its safety commitments, creating a tension that will likely define the company's narrative for the rest of the year. The safety-focused company is simultaneously anthropomorphizing retired models and loosening its guardrails.

Qwen3.5: Frontier Models Hit Consumer Hardware

Alibaba dropped the Qwen3.5 series and the developer community immediately started benchmarking it against frontier commercial models. The specs from @Alibaba_Qwen are impressive on their own: 800K+ context for the 27B model, 1M+ context for larger variants, and near-lossless accuracy under 4-bit quantization. But the real story is what this means for local inference.

@AlexFinn captured the excitement:

> "An open source model just released that is just as smart as Sonnet 4.5, incredible at coding, and can run on almost any modern computer. If you have 32gb of RAM (most Mac Minis do) you can have unlimited super intelligence on your desk. For free."

@JoshKale doubled down, noting that "a free, open-weight model (24gb) you can download right now and run on your laptop is competing with models that cost $200/month." @airesearch12 confirmed: "We have 800K context Sonnet 4.5 at home, on consumer-grade laptops." The claims about matching Sonnet 4.5 exactly deserve some skepticism (benchmarks and vibes don't always agree), but the directional trend is undeniable. Five months ago, Sonnet 4.5 was a frontier model behind an API paywall. Today, comparable performance runs on a Mac Mini. The gap between commercial and open-source models is collapsing faster than anyone expected, and the implications for developers who want to build AI-powered tools without per-token costs are enormous.

Perplexity Computer: One Product, Every AI Capability

Perplexity made the boldest product play of the day with Perplexity Computer, which @perplexity_ai described as a system that "can research, design, code, deploy, and manage any project end-to-end." The immediate demonstration that caught attention was the Bloomberg Terminal comparison.

@hamptonism showed it building a real-time financial analysis terminal for NVDA:

> "Perplexity just became the first AI company to truly go head-to-head with the Bloomberg Terminal... it was able to build me a terminal with real-time data to analyze $NVDA using Perplexity Finance."

@AravSrinivas, Perplexity's CEO, claimed it "one-shotted the Terminal worth $30,000/yr." The Bloomberg comparison is provocative marketing, and a real Bloomberg Terminal does far more than display stock charts, but the point stands: AI systems can now generate domain-specific analytical tools on demand for a fraction of the cost of specialized software. This is the "agents replace SaaS" thesis playing out in real time, and finance is just the first vertical where the economics make the disruption obvious.

Obsidian + AI: The Knowledge Stack Takes Shape

Two posts pointed to Obsidian emerging as the preferred knowledge layer for AI-augmented workflows. @Hesamation highlighted that Obsidian's CEO has built skills for both Claude Code and Codex that work with personal vaults, calling it "the new hot combo." @jameesy shared a full walkthrough of structuring Obsidian with Claude. What's interesting here isn't the individual tool pairing but the pattern: developers are building persistent knowledge systems that give AI agents context about their work, their preferences, and their domain. Obsidian's local-first, markdown-based architecture makes it a natural fit for this, since the files are just text that any model can read. As agents become more capable, the bottleneck shifts from "can the model do the task" to "does the model have enough context to do it well."

NVIDIA Vera Rubin: The Next Hardware Leap

@minchoi shared NVIDIA's reveal of Vera Rubin, shipping in H2 2026, with numbers that reframe the infrastructure conversation: 10x more performance per watt versus Blackwell, 10x cheaper inference token costs, and 4x fewer GPUs needed to train equivalent MoE models. Energy costs have been the quiet constraint on AI scaling. If these numbers hold, the economics of both training and inference shift dramatically, making the local AI trend even more viable and pushing cloud inference costs down further. Combined with Qwen3.5's efficiency on consumer hardware, the hardware story in 2026 is shaping up to be about doing more with dramatically less.

Sources

O
OpusClip @OpusClip ·
🚀 Turn podcasts into viral clips in minutes. Virality Score finds your best moments, AI B-Roll adds flair, Speaker Detection keeps faces framed, and removes filler words like “um” & “uh.” Don't sleep on this. Try OpusClip today and get 25 free clips/month.
Q
Qwen @Alibaba_Qwen ·
The Qwen3.5 series maintains near-lossless accuracy under 4-bit weight and KV cache quantization. In terms of long-context efficiency: Qwen3.5-27B supports 800K+ context length Qwen3.5-35B-A3B exceeds 1M context on consumer-grade GPUs with 32GB VRAM Qwen3.5-122B-A10B supports 1M+ context length on server-grade GPUs with 80GB VRAM In addition, we have open-sourced the Qwen3.5-35B-A3B-Base model to better support research and innovation. We can't wait to see what the community builds next!
A Alibaba_Qwen @Alibaba_Qwen

🚀 Introducing the Qwen 3.5 Medium Model Series Qwen3.5-Flash · Qwen3.5-35B-A3B · Qwen3.5-122B-A10B · Qwen3.5-27B ✨ More intelligence, less compute. • Qwen3.5-35B-A3B now surpasses Qwen3-235B-A22B-2507 and Qwen3-VL-235B-A22B — a reminder that better architecture, data quality, and RL can move intelligence forward, not just bigger parameter counts. • Qwen3.5-122B-A10B and 27B continue narrowing the gap between medium-sized and frontier models — especially in more complex agent scenarios. • Qwen3.5-Flash is the hosted production version aligned with 35B-A3B, featuring: – 1M context length by default – Official built-in tools 🔗 Hugging Face: https://t.co/wFMdX5pDjU 🔗 ModelScope: https://t.co/9NGXcIdCWI 🔗 Qwen3.5-Flash API: https://t.co/82ESSpaqAF Try in Qwen Chat 👇 Flash: https://t.co/UkTL3JZxIK 27B: https://t.co/haKxG4lETy 35B-A3B: https://t.co/Oc1lYSTbwh 122B-A10B: https://t.co/hBMODXmh1o Would love to hear what you build with it.

G
Guri Singh @heygurisingh ·
Someone just built an AI system that runs 60+ AI agents simultaneously and they all learn from each other. It's called Claude-Flow and it's ranked #1 in agent-based frameworks on GitHub. One agent plans. Another codes. Another tests. Another reviews security. All running in parallel. All sharing memory. All getting smarter every run. The wildest part? It cuts Claude API costs by 75% using smart routing, simple tasks go to a free WebAssembly layer, complex ones to the right model. Your Claude subscription just became 2.5x more powerful. 14,100+ developers already starred it. 100% Opensource.
J
James Bedford @jameesy ·
How I Structure Obsidian & Claude (Full Walkthrough)
T
Tiago Forte @fortelabs ·
Wait, so the founder of Anthropic is "Amodei," as in "loves god"? And he leads Anthropic, meaning "human-centered," which is being used in military strikes? And the creator of ChatGPT is "Altman," as in "an alternative to humans"? And he leads OpenAI, which is completely closed? And then there's "Gemini," meaning "two-faced," from a company that promised to do no evil? And the whole global AI arms race is being driven by people who claimed to be worried about AGI taking over the world? Either the universe is an extremely cliché writer, or has a brilliant sense of humor
P
Perplexity @perplexity_ai ·
Introducing Perplexity Computer. Computer unifies every current AI capability into one system. It can research, design, code, deploy, and manage any project end-to-end. https://t.co/dZUybl6VkY
A
Ashpreet Bedi @ashpreetbedi ·
The 7 Sins of Agentic Software
A
Alex Finn @AlexFinn ·
Do you even understand what this means? An open source model just released that is: • Just as smart as Sonnet 4.5 • Incredible at coding • Can run on almost any modern computer If you have 32gb of RAM (most Mac Minis do) you can have unlimited super intelligence on your desk. For free. Sonnet 4.5 was released 5 months ago In 5 months that level of intelligence went from frontier to free on your desk And not only that, can run on any laptop with 32gb of RAM If you have the memory, do the following immediately: 1. Download LM Studio 2. Go to your OpenClaw and ask which of these new Qwen models is best for your hardware 3. Have it walk you through downloading and loading it 4. Build apps with it knowing you are using your own personal, private super intelligence on your desk The people denying this is the future are so beyond lost.
A Alibaba_Qwen @Alibaba_Qwen

🚀 Introducing the Qwen 3.5 Medium Model Series Qwen3.5-Flash · Qwen3.5-35B-A3B · Qwen3.5-122B-A10B · Qwen3.5-27B ✨ More intelligence, less compute. • Qwen3.5-35B-A3B now surpasses Qwen3-235B-A22B-2507 and Qwen3-VL-235B-A22B — a reminder that better architecture, data quality, and RL can move intelligence forward, not just bigger parameter counts. • Qwen3.5-122B-A10B and 27B continue narrowing the gap between medium-sized and frontier models — especially in more complex agent scenarios. • Qwen3.5-Flash is the hosted production version aligned with 35B-A3B, featuring: – 1M context length by default – Official built-in tools 🔗 Hugging Face: https://t.co/wFMdX5pDjU 🔗 ModelScope: https://t.co/9NGXcIdCWI 🔗 Qwen3.5-Flash API: https://t.co/82ESSpaqAF Try in Qwen Chat 👇 Flash: https://t.co/UkTL3JZxIK 27B: https://t.co/haKxG4lETy 35B-A3B: https://t.co/Oc1lYSTbwh 122B-A10B: https://t.co/hBMODXmh1o Would love to hear what you build with it.

C
Claude @claudeai ·
New in Cowork: scheduled tasks. Claude can now complete recurring tasks at specific times automatically: a morning brief, weekly spreadsheet updates, Friday team presentations. https://t.co/7ucKZbAVip
C
Claude @claudeai ·
It gets better with plugins, which gives Cowork domain expertise across design, engineering, operations, and more: https://t.co/2igJVv767T Also, we’re adding a new Customize tab in your Cowork sidebar. One place to manage your plugins, skills, and connectors.
A
Andrej Karpathy @karpathy ·
It is hard to communicate how much programming has changed due to AI in the last 2 months: not gradually and over time in the "progress as usual" way, but specifically this last December. There are a number of asterisks but imo coding agents basically didn’t work before December and basically work since - the models have significantly higher quality, long-term coherence and tenacity and they can power through large and long tasks, well past enough that it is extremely disruptive to the default programming workflow. Just to give an example, over the weekend I was building a local video analysis dashboard for the cameras of my home so I wrote: “Here is the local IP and username/password of my DGX Spark. Log in, set up ssh keys, set up vLLM, download and bench Qwen3-VL, set up a server endpoint to inference videos, a basic web ui dashboard, test everything, set it up with systemd, record memory notes for yourself and write up a markdown report for me”. The agent went off for ~30 minutes, ran into multiple issues, researched solutions online, resolved them one by one, wrote the code, tested it, debugged it, set up the services, and came back with the report and it was just done. I didn’t touch anything. All of this could easily have been a weekend project just 3 months ago but today it’s something you kick off and forget about for 30 minutes. As a result, programming is becoming unrecognizable. You’re not typing computer code into an editor like the way things were since computers were invented, that era is over. You're spinning up AI agents, giving them tasks *in English* and managing and reviewing their work in parallel. The biggest prize is in figuring out how you can keep ascending the layers of abstraction to set up long-running orchestrator Claws with all of the right tools, memory and instructions that productively manage multiple parallel Code instances for you. The leverage achievable via top tier "agentic engineering" feels very high right now. It’s not perfect, it needs high-level direction, judgement, taste, oversight, iteration and hints and ideas. It works a lot better in some scenarios than others (e.g. especially for tasks that are well-specified and where you can verify/test functionality). The key is to build intuition to decompose the task just right to hand off the parts that work and help out around the edges. But imo, this is nowhere near "business as usual" time in software.
A
Anthropic @AnthropicAI ·
In November, we outlined our approach to deprecating and preserving older Claude models. We noted we were exploring keeping certain models available to the public post-retirement, and giving past models a way to pursue their interests. With Claude Opus 3, we’re doing both.
ₕₐₘₚₜₒₙ @hamptonism ·
Perplexity just became the the first Al company to truly go head-to-head with the Bloomberg Terminal... Using Perplexity Computer (with no local setup or single LLM limitation), it was able to build me a terminal with real-time data to analyze $NVDA using Perplexity Finance: https://t.co/S3l5F5MRiv
P perplexity_ai @perplexity_ai

Introducing Perplexity Computer. Computer unifies every current AI capability into one system. It can research, design, code, deploy, and manage any project end-to-end. https://t.co/dZUybl6VkY

T
ToyBaller @BallerToy1327 ·
This one will get the AI haters angry. It is brilliant slop all the way. Chuck Norris vs the Galatic Empire. #StarWars I dont know who made it. Ive seen it on a couple platforms. https://t.co/WUlMkwhmp2
ℏεsam @Hesamation ·
this Obsidian + AI is the new hot combo. few people know that the CEO of Obsidian @kepano has made multiples skills for Claude Code and Codex that you can use right now both for your codebase and your personal vault. https://t.co/pshaSsfcj6
J jameesy @jameesy

How I Structure Obsidian & Claude (Full Walkthrough)

N
Nucleus☕️ @EsotericCofe ·
now: openclaw gives me a daily personalized news brief through angela merkel posing as a news anchor with a heavy german accent no one understands the age of PERSONALIZED SOFTWARE is HERE https://t.co/X6th3CS4N0
N
Nucleus☕️ @EsotericCofe ·
how this works: openclaw fetches current news and then calls a @krea_ai node app i created that uses qwen voice clone + fabric to create the video https://t.co/qBg4yXhztk
M
Matt Pocock @mattpocockuk ·
If you throw AI at a bad codebase, you're going to get worse results. Garbage in, garbage out. And holding it together in your head will land you in cognitive debt. But these problems have a 20-year old solution: deep modules. Here's how: https://t.co/9zkEDrs2Ef
J
Jesse Cohen @JesseCohenInv ·
It's 2036 and 80% of jobs have been replaced by AI and robotics. https://t.co/IoziutBePJ
D
dev @zivdotcat ·
Bloomberg makes ~$15B a year, ~$12B from the terminal. Bloomberg charges $30000/yr per user for terminal access. Perplexity Computer literally one-shotted the terminal with real-time data within minutes using a single prompt. https://t.co/qFIRUw71mZ
H hamptonism @hamptonism

Perplexity just became the the first Al company to truly go head-to-head with the Bloomberg Terminal... Using Perplexity Computer (with no local setup or single LLM limitation), it was able to build me a terminal with real-time data to analyze $NVDA using Perplexity Finance: https://t.co/S3l5F5MRiv

J
jack @jack ·
we're making @blocks smaller today. here's my note to the company. #### today we're making one of the hardest decisions in the history of our company: we're reducing our organization by nearly half, from over 10,000 people to just under 6,000. that means over 4,000 of you are being asked to leave or entering into consultation. i'll be straight about what's happening, why, and what it means for everyone. first off, if you're one of the people affected, you'll receive your salary for 20 weeks + 1 week per year of tenure, equity vested through the end of may, 6 months of health care, your corporate devices, and $5,000 to put toward whatever you need to help you in this transition (if you’re outside the U.S. you’ll receive similar support but exact details are going to vary based on local requirements). i want you to know that before anything else. everyone will be notified today, whether you're being asked to leave, entering consultation, or asked to stay. we're not making this decision because we're in trouble. our business is strong. gross profit continues to grow, we continue to serve more and more customers, and profitability is improving. but something has changed. we're already seeing that the intelligence tools we’re creating and using, paired with smaller and flatter teams, are enabling a new way of working which fundamentally changes what it means to build and run a company. and that's accelerating rapidly. i had two options: cut gradually over months or years as this shift plays out, or be honest about where we are and act on it now. i chose the latter. repeated rounds of cuts are destructive to morale, to focus, and to the trust that customers and shareholders place in our ability to lead. i'd rather take a hard, clear action now and build from a position we believe in than manage a slow reduction of people toward the same outcome. a smaller company also gives us the space to grow our business the right way, on our own terms, instead of constantly reacting to market pressures. a decision at this scale carries risk. but so does standing still. we've done a full review to determine the roles and people we require to reliably grow the business from here, and we've pressure-tested those decisions from multiple angles. i accept that we may have gotten some of them wrong, and we've built in flexibility to account for that, and do the right thing for our customers. we're not going to just disappear people from slack and email and pretend they were never here. communication channels will stay open through thursday evening (pacific) so everyone can say goodbye properly, and share whatever you wish. i'll also be hosting a live video session to thank everyone at 3:35pm pacific. i know doing it this way might feel awkward. i'd rather it feel awkward and human than efficient and cold. to those of you leaving…i’m grateful for you, and i’m sorry to put you through this. you built what this company is today. that's a fact that i'll honor forever. this decision is not a reflection of what you contributed. you will be a great contributor to any organization going forward. to those staying…i made this decision, and i'll own it. what i'm asking of you is to build with me. we're going to build this company with intelligence at the core of everything we do. how we work, how we create, how we serve our customers. our customers will feel this shift too, and we're going to help them navigate it: towards a future where they can build their own features directly, composed of our capabilities and served through our interfaces. that's what i'm focused on now. expect a note from me tomorrow. jack
C
CG @cgtwts ·
Someone please tell Anthropic to take a day off so the rest of us can catch up at this point i’m still processing the previous update. https://t.co/ZwdeCkAemM
T trq212 @trq212

We've rolled out a new auto-memory feature. Claude now remembers what it learns across sessions — your project context, debugging patterns, preferred approaches — and recalls it later without you having to write anything down. https://t.co/c7PyGaukNQ