AI Learning Digest.

Seedance 3.0 Leak Promises 18-Minute AI Films as Google Fires Back with Veo 4

Daily Wrap-Up

If you looked at the AI timeline on February 14th and squinted, you'd think Hollywood had already lost. Seedance was everywhere. Not just the shipping 2.0 release, which had people generating anime-to-live-action conversions and meme videos with startling fidelity, but leaked specs for a 3.0 version that reads like science fiction: 18-minute coherent films, persistent narrative memory across scenes, native multilingual emotional dubbing. Whether those specs are real or aspirational marketing, they captured something true about the moment. The gap between "AI can make a cool 15-second clip" and "AI can make something you'd actually watch" is closing faster than anyone expected. Google apparently agreed, with reports of Veo 4 dropping as a direct response.

On the developer tooling side, Claude Code continued its quiet evolution into an ecosystem. Session persistence daemons, Tamagotchi visualizations, lazy-loaded MCP tools. The interesting thread wasn't any single tool but what @addyosmani surfaced from Claude Code's creator Boris: that the engineer's value is shifting from code generation to judgment, taste, and systems thinking. That framing landed alongside @thdxr's much blunter reality check about what AI adoption actually looks like inside large organizations: unmotivated teams churning out slop, CFOs shocked by per-engineer LLM bills, and the two people who actually tried now drowning in low-quality output. The tension between "AI makes everything possible" and "most organizations can't absorb that possibility" was the real undercurrent of the day.

The most practical takeaway for developers: if you're building tools or products on top of AI, pay attention to the WebMCP spec that Google and Microsoft co-authored. It turns websites into structured tool APIs for agents, replacing brittle screenshot-and-click approaches with clean function calls. Early benchmarks show 67% less compute and 98% task accuracy. If agent-driven traffic becomes significant, sites that expose WebMCP endpoints will get preferential treatment from agents, the same way structured data won the SEO game. Start thinking about what your "agent experience" looks like.

Quick Hits

  • @jxmnop highlighted Andrej Karpathy implementing most of modern LLM complexity in under 200 lines with just three imports. The kind of educational artifact that makes dense papers click.
  • @DannyLimanseta shared a clever workflow: using Google Gemini AI Studio to build a custom art generation tool for consistent game assets, then manually adjusting proportions in Photoshop. The hybrid AI-plus-manual approach that actually ships.
  • @DannyLimanseta also showed off procedurally generated Diablo-style items with affixes and rarities, built with a vibe-coded custom art generation tool. Mesmerizing stuff.
  • @sachinyadav699 posted the inevitable cycle meme: "Hard times create strong men. Strong men create C. C creates good times. Good times create Python programmers. Python programmers create AI. AI creates vibe coders. Vibe coders create weak men."
  • @PeterMeijer asked Claude to plan a military operation to capture the President of Venezuela. The kind of prompt that makes safety teams lose sleep.
  • @VicVijayakumar's reaction to something unnamed: "yeah yeah this sounds pretty normal... [keeps reading] wait what." Context-free intrigue at its finest.
  • @theCTO on someone named Dax: "dax chose war. against literally everyone. hell yes." No further context provided or needed.
  • @Babygravy9 declared "at last, the AI meme we all wanted has been made." The bar for AI-generated memes apparently has been met.

Seedance Dominates: AI Video's Breakout Moment

ByteDance's Seedance owned the timeline with a one-two punch: a genuinely impressive 2.0 release generating real engagement, and leaked 3.0 specs that set imaginations on fire. The 2.0 release alone had the community producing everything from anime-to-live-action conversions to action sequences mixing Neo, John Wick, and the Terminator. @VraserX captured the excitement: "Seedance 2.0 basically means we're about to make live action versions of our favorite anime ourselves. Fan fiction just evolved into fan cinema." The character replacement feature proved particularly compelling, with @markgadala noting it would "lead to a ton of these videos."

But it was the 3.0 leak that truly captured attention. @mark_k provided the most detailed breakdown, citing Chinese social media sources claiming the next version supports "seamless single-take generations of 10+ minutes" with internal tests reaching 18 minutes, native multilingual emotional dubbing, Hollywood-grade director controls accepting storyboard scripts and real-time shot commands, and compute costs at one-eighth of the 2.0 version. @VraserX distilled it more bluntly: "Hollywood's moat was scale, capital, and distribution. AI just compressed all three."

The competitive response was immediate. @markgadala reported Google answering with Veo 4, calling it "even better" and predicting "a wild year." @cfryant observed that a single AI video "brought all the fence sitters to the pro AI side," while @MarvelLatin's Spanish-language plea that Hollywood must stop AI's advance reflected genuine anxiety from the creative industry. What makes this moment different from previous AI video hype cycles is the sheer volume of people actually using the tools and sharing results, not just reacting to demos. @Dheepanratnam praised the technical achievement of rendering speed without background blur artifacts, noting the prompt engineering around these tools is becoming "basically a textbook." Whether the 3.0 specs are real or aspirational, the 2.0 output speaks for itself: we've crossed from "interesting tech demo" to "people are making things they want to watch."

Claude Code's Ecosystem Matures

Claude Code's growth story on this day wasn't about a single feature announcement but about an ecosystem crystallizing around it. The tooling layer is getting surprisingly deep. @AdamTzag built a launchd daemon that watches for running Claude Code sessions and saves them when Ghostty exits, restoring everything on next launch: "Basically pgrep + sleep in a loop. 2MB of memory doing nothing until you quit." Meanwhile, @SamuelBeek discovered someone had built a Claude Code Tamagotchi, prompting @chloevdl014 to reply: "a tamagotchi for your coding agent is genuinely the best idea I've seen this week. need this to guilt trip me when I ignore its suggestions."

On the technical side, @bcherny (who works on Claude Code at Anthropic) shared that the tool "intelligently loads MCP tools on demand," lazy-loading when there are many tools and loading more upfront when there are few. This kind of invisible optimization matters as MCP tool counts grow. @alexhillman teased something as "legitimately the most interesting thing to hit Claude Code desktop" without elaborating.

The deeper thread came from @addyosmani, who highlighted Claude Code creator Boris's framing that AI shifts the engineer's value to "what do we build, why, for whom, and how it all fits together. The bottleneck was always judgment, taste, and systems thinking. AI just made that more obvious." @bcherny reinforced this in a separate reply: "Someone has to prompt the Claudes, talk to customers, coordinate with other teams, decide what to build next." The message is consistent: coding agents don't eliminate engineering, they surface what engineering was always supposed to be about.

Enterprise Reality Checks

For every breathless AI capability post, there was a corresponding reality check about what adoption actually looks like at scale. @thdxr dropped the day's most grounding thread about AI in large engineering organizations. The picture isn't pretty: "your org rarely has good ideas. ideas being expensive to implement was actually helping." The majority of workers use AI "to churn out their tasks with less energy spend" rather than to be more effective, while "the 2 people on your team that actually tried are now flattened by the slop code everyone is producing, they will quit soon." And then there's the CFO asking why each engineer now costs $2,000 extra per month in LLM bills.

That cost concern is already driving behavior. @thdxr followed up noting that a 20,000-developer company is "looking at these numbers and going W T F and they're moving inference to their own GPU cluster with open source models. There isn't infinite budget and appetite for this stuff." @staysaasy offered a different angle on the gap between AI capability and real-world impact: "New iOS apps have exploded in the last six months due to AI coding. Number of new apps people have recommended to me: 0."

The most forward-looking enterprise post came from @aakashgupta on WebMCP, the spec Google and Microsoft co-authored that turns websites into structured APIs for AI agents. Instead of agents taking screenshots and guessing at buttons, sites register structured tools via navigator.modelContext. Early benchmarks show 67% less compute overhead and 98% task accuracy. The second-order effect is what matters: "the site that exposes structured tools gives the agent a clean, reliable path. The site that doesn't forces the agent to fumble through the UI. Agents will prefer the cheaper path. Every time." Agent Experience Optimization is about to become a real discipline.

Models and the Local Inference Push

The push toward local AI inference continued gaining momentum, driven by both improving open-source models and enterprise cost pressure. @himanshustwts shared early impressions of GLM-5 running through Claude Code, calling it "impressively good in design (one-shotted better UI than Opus 4.6)," noting it's "actually not sycophantic," has "nearly no hallucinations," and is "way more optimized for coding." The conclusion: "Chinese competition is drilling the perf up. Models are times cheaper than Opus. We all win."

@meta_alchemist laid out a vision for the local AI endgame that reads like a homelab manifesto: start with Claude or OpenAI, move to local inference on your own hardware, upgrade with income from your builds, run agents 24/7, and "earn passively via your agents." The prerequisite, they argue, is that "open source local models like Minimax 2.5 already reached Opus 4.5 levels." Whether that benchmark claim holds up to scrutiny, the trajectory is clear. Combined with @thdxr's report of a 20,000-dev company moving to self-hosted open-source models, the economics of hosted inference are creating real gravitational pull toward local deployment. The question isn't whether local models are good enough; it's whether the operational overhead of running your own GPU cluster is worth the cost savings. For a 20,000-person org spending $2,000 per developer per month on API calls, the math gets obvious fast.

Source Posts

s
staysaasy @staysaasy ·
New iOS apps have exploded in the last six months due to AI coding. Number of new apps people have recommended to me: 0. Number of new apps people have mentioned to me that they’re using: 0. Number of new apps that I’ve downloaded: 1.
B
Boris Cherny @bcherny ·
@big_duca Someone has to prompt the Claudes, talk to customers, coordinate with other teams, decide what to build next. Engineering is changing and great engineers are more important than ever.
A
Adam Tzagournis, CPA @AdamTzag ·
Every time I quit Ghostty or restart my Mac I'd lose all my Claude Code sessions. 10 tabs across different conversations etc. So I wrote a launchd daemon that watches for running sessions and saves them when Ghostty exits. Next time you open Ghostty everything comes back, same tabs, same conversations, same CLI flags. Basically pgrep + sleep in a loop. 2MB of memory doing nothing until you quit. https://t.co/OXczQT7cSv Think you'll like this @bcherny @trq212 @mitchellh
V
VraserX e/acc @VraserX ·
Seedance 3.0 specs just leaked. If this is accurate, this isn’t another incremental AI video upgrade. It’s a structural shock to Hollywood. • 10 to 18 minute coherent films in one pass • Persistent narrative memory across scenes • Native multi language voice with emotional control • Shot level directing inputs like a real production workflow • Cost reportedly a fraction of traditional shoots We are not talking about TikTok clips anymore. We are talking about full cinematic episodes generated from prompts. Hollywood’s moat was scale, capital, and distribution. AI just compressed all three.
刘正博士 @mokoocn

Seedance 3.0已进入闭门冲刺阶段,并实现多项颠覆性技术跃迁! 这一代不再满足于15秒短片,而是直接把AI视频生成推向“长篇电影时代”,让任何人用一句话就能产出带完整剧情、多镜头转场、原生多声道配音的10分钟+商业级内容! 据多位接近项目核心的消息源透露,Seedance 3.0的核心杀招包括: 1,无限时长连续生成:突破现有模型的长度瓶颈,支持单次生成最长10分钟以上无缝视频(内部测试已达18分钟无明显崩坏),通过全新“叙事记忆链”架构,AI能记住前文剧情、角色性格、场景设定,自动规划多幕结构、悬念铺垫和高潮转折,像真人导演一样“讲故事”! 2,原生多语言+情绪配音同步:不再是后期配音,而是端到端联合训练,生成视频时同时输出自然唇形同步的中文、英文、日语、韩语等多语种对白,甚至能根据角色情绪自动调整语调、呼吸、哭腔、笑声。测试片段中,AI生成的武侠片人物对白已达到专业配音演员水准! 3,电影级可控导演工具:支持“分镜脚本输入”+“实时导演指令”,用户可直接写“镜头1:广角推轨,英雄从废墟中起身;镜头2:快速剪辑追车戏,配重低音鼓点”,AI瞬间理解并执行。还内置行业标准色调预设(IMAX、胶片风、Netflix调色等),一键出片即可送审! 4,超低成本核弹:得益于新一代蒸馏+高效推理优化,生成1分钟电影级视频的算力成本已降至Seedance 2.0的1/8,相当于传统剧组单场戏的几百分之一。独立导演、短剧公司、广告主将迎来史诗级降维打击!

📙
📙 Alex Hillman @alexhillman ·
This is legitimatly the most interesting thing to hit Claude Code desktop 👀
A Anthony Morris ツ @amorriscode

SSH support is now available for Claude Code on desktop Connect to your remote machines and let Claude cook, TMUX optional. https://t.co/sVtMSROjRu

V
Vidal Chloé @chloevdl014 ·
@SamuelBeek ok a tamagotchi for your coding agent is genuinely the best idea i've seen this week. need this to guilt trip me when I ignore its suggestions lol
d
dax @thdxr ·
everyone's talking about their teams like they were at the peak of efficiency and bottlenecked by ability to produce code here's what things actually look like - your org rarely has good ideas. ideas being expensive to implement was actually helping - majority of workers have no reason to be super motivated, they want to do their 9-5 and get back to their life - they're not using AI to be 10x more effective they're using it to churn out their tasks with less energy spend - the 2 people on your team that actually tried are now flattened by the slop code everyone is producing, they will quit soon - even when you produce work faster you're still bottlenecked by bureaucracy and the dozen other realities of shipping something real - your CFO is like what do you mean each engineer now costs $2000 extra per month in LLM bills
A
Aakash Gupta @aakashgupta ·
Google and Microsoft just co-authored the spec that turns every website into an API for AI agents. The second-order effects here are massive. Right now, browser agents work by taking screenshots, parsing the DOM, and guessing which buttons to click. It works about as well as you’d expect. Fragile, expensive, slow. WebMCP replaces all of that with a single browser API: navigator.modelContext. Websites register structured tools directly in client-side JavaScript. The agent reads a menu of available actions, calls them, gets structured data back. No scraping. No backend MCP server in Python or Node. The tools run inside the browser tab and share the user’s existing auth session. Early benchmarks show ~67% reduction in computational overhead compared to visual agent-browser interactions. Task accuracy around 98%. The second-order effect is where this gets wild. Today, when a browser agent visits two competing airline sites, it’s guessing at both interfaces equally. Once WebMCP adoption spreads, the site that exposes structured tools gives the agent a clean, reliable path to complete the task. The site that doesn’t forces the agent to fumble through the UI. Agents will prefer the cheaper path. Every time. This means “Agent Experience Optimization” becomes a real discipline. Tool naming, schema design, description quality. Sound familiar? It’s the same shift that happened when meta descriptions and structured data became optimization surfaces for search engines. Except this time, the traffic source isn’t Google’s crawler. It’s every AI agent on the internet. Bots already make up 51% of web traffic. Google just gave them a front door.
C Chrome for Developers @ChromiumDev

WebMCP is available for early preview → https://t.co/bZMcANfg37 WebMCP aims to provide a standard way for exposing structured tools, ensuring AI agents can perform actions on your side with increased speed, reliability, and precision. https://t.co/9NvSi6rMdV

d
dr. jack morris @jxmnop ·
this is really amazing. Andrej implemented most of the complexity of modern LLMs in <200 lines. and only imported these three packages https://t.co/nihmfI2TN0
A Andrej Karpathy @karpathy

New art project. Train and inference GPT in 243 lines of pure, dependency-free Python. This is the *full* algorithmic content of what is needed. Everything else is just for efficiency. I cannot simplify this any further. https://t.co/HmiRrQugnP

h
himanshu @himanshustwts ·
Dude this is crazy model. been quite sometime i have been vibe testing GLM-5 on Claude Code. My first initial impressions: + impressively good in design (one shotted better ui with glm than Opus 4.6) + it is actually not sycophantic (tried rigorously) + nearly no hallucinations. + model is way more optimized for coding. i am loving the game. chinese competition is drilling the perf up. models are times cheaper than Opus. we all win.
Z Z.ai @Zai_org

Introducing GLM-5: From Vibe Coding to Agentic Engineering GLM-5 is built for complex systems engineering and long-horizon agentic tasks. Compared to GLM-4.5, it scales from 355B params (32B active) to 744B (40B active), with pre-training data growing from 23T to 28.5T tokens. Try it now: https://t.co/WCqWT0raFJ Weights: https://t.co/DteNDHjSEh Tech Blog: https://t.co/Wxn5ARTJxH OpenRouter (Previously Pony Alpha): https://t.co/7Khf64Lxg6 Rolling out from Coding Plan Max users: https://t.co/Nk8Y98Il7s

V
Vic 🌮 @VicVijayakumar ·
yeah yeah this sounds pretty normal. everyone is doing this, you're not special [keeps reading] wait what
p patagucci perf papi @kenwheeler

i feel like a lot of people i know aren’t as tapped in on ai setups as they should be. lots of butt sniffing, curious pokes. i’m gonna tell you exactly how i have this shit set up, and it may be dumb, but it’s a set up:

D
Danny Limanseta @DannyLimanseta ·
Thanks! Here's what I did: - I used Google Gemini AI Studio to whip up an art gen tool to create my art assets in a specific style - I used Google Nano Banana pro API to generate the images, using some images as style references - It's able to generate quite consistent art assets but I still had to do some manual photoshop to adjust proportions to make it fit into my game's customisation system.
a
adam @theCTO ·
dax chose war. against literally everyone. hell yes
d dax @thdxr

everyone's talking about their teams like they were at the peak of efficiency and bottlenecked by ability to produce code here's what things actually look like - your org rarely has good ideas. ideas being expensive to implement was actually helping - majority of workers have no reason to be super motivated, they want to do their 9-5 and get back to their life - they're not using AI to be 10x more effective they're using it to churn out their tasks with less energy spend - the 2 people on your team that actually tried are now flattened by the slop code everyone is producing, they will quit soon - even when you produce work faster you're still bottlenecked by bureaucracy and the dozen other realities of shipping something real - your CFO is like what do you mean each engineer now costs $2000 extra per month in LLM bills

s
sam @SamuelBeek ·
someone made a Claude Code Tamagochi with https://t.co/YEnlZTf04c whuuutt, absolutely wild, this tool didn't exist 1 week ago https://t.co/Nd85Yk4msb
d
dax @thdxr ·
btw we spoke to a company yesterday that's at the scale of 20,000 devs they are looking at these numbers and going W T F and they're moving inference to their own gpu cluster with open source models there isn't infinite budget and appetite for this stuff
M
Mark Kretschmann @mark_k ·
Rumors about Seedance 3.0 have surfaced on Chinese X, and they are mind-blowing if true: According to leaks from Dr. Liu Zheng (@mokoocn), ByteDance's video generation AI has entered its final "closed-door sprint" and is aiming to end the era of short clips forever. If these specs are real, we are looking at the "feature film era" of AI: 🎥 **Infinite Continuous Generation** The 15-second limit is dead. Seedance 3.0 reportedly supports seamless single-take generations of **10+ minutes** (with internal tests reaching 18 minutes without collapse). It uses a "narrative memory chain" to remember plot points, character personalities, and settings, effectively allowing it to "direct" multi-act stories with suspense and twists like a human. 🗣️ **Native Multi-Language & Emotional Dubbing** No more post-production lip-syncing. The model allegedly handles end-to-end video *and* audio generation simultaneously. It can output perfect lip-sync in Chinese, English, Japanese, and Korean, while dynamically adjusting breathing, crying, or laughing to match the character's emotional state. 🎬 **Hollywood-Grade Director Control** Forget simple prompts; this supports "storyboard script input" and real-time director commands (e.g., "Shot 1: Wide-angle dolly push..."). It reportedly understands cinematic language instantly and includes industry-standard color grading presets like IMAX and Netflix-style looks. 📉 **The "Nuclear Bomb" Cost Reduction** Perhaps the most disruptive claim: The compute cost for 1 minute of cinematic video has supposedly dropped to **1/8th of Seedance 2.0**. This would make high-end video production cost pennies compared to traditional crews, described as a "dimension-reduction strike" against the ad and short-drama industries. If this ships, the barrier to entry for creating full-length movies just evaporated.
M
Meta Alchemist @meta_alchemist ·
Very soon PC parts will skyrocket imo and people will have to wait very long to get what they want as most businesses too will rush for local LLMs due to privacy and cost efficiency, since open source local models like Minimax 2.5 already reached Opus 4.5 levels Endgame is looking like this: > start with claude / openai > openclaw with your local pc or mac > upgrade your hardware with income from your builds and products > go private > run more agents 24/7 > let them build autonomously > earn passively via your agents > build things that you love > spend more time with your loved ones This is the road to riches in the age of AI as long as you put in the time to master vibe coding, ai tools, and distribution You'll need this:
M Meta Alchemist @meta_alchemist

Best PC Specs to Run Local AI Models like Minimax, Free!

M
Meta Alchemist @meta_alchemist ·
Best PC Specs to Run Local AI Models like Minimax, Free!
P
Peter Meijer @PeterMeijer ·
Ok Claude, tell me how to capture the President of Venezuela without any of our guys dying. Make no mistakes. https://t.co/SYtibYUucU
T The Wall Street Journal @WSJ

Breaking: The Pentagon used Anthropic’s AI tool Claude in its military operation to capture former Venezuelan President Nicolás Maduro https://t.co/HwXRnafUo0

C
Christopher Fryant @cfryant ·
This one AI video just brought all the fence sitters to the pro AI side.
C Charles Curran @charliebcurran

Seedance 2.0 Prompt: Sum up the AI discourse in a meme - make sure it’s retarded and gets 50 likes. https://t.co/09yPdo3Tjy

B
Boris Cherny @bcherny ·
@nikil 👋 Claude Code intelligently loads MCP tools on demand — if you have lots of tools they will be lazy-loaded, and if just a few it will load more upfront. No need to optimize this yourself, Claude will take care of it for you.