AI Learning Digest.

Factory AI Drops Agent Readiness Framework While Claude Code's Skills Ecosystem Goes Mainstream

Daily Wrap-Up

The throughline today is infrastructure. Not the servers-and-containers kind, but the organizational kind. The question has shifted from "can AI agents write code?" to "is your codebase ready for them to work in?" Factory AI's Agent Readiness framework gave that question a formal shape, and the response was immediate. Engineering leaders chimed in that this should be priority one. Meanwhile, the Claude Code community is building out the connective tissue that makes agents actually useful: visual feedback tools, skills marketplaces, date-time injection hooks, and CPU profiling output formatted for LLMs. It's plumbing season, and the people laying pipe are winning.

The most surprising moment came from @GergelyOrosz, who noted that inside Big Tech, the internal token usage leaderboards are dominated by distinguished engineers and even VPs. Not junior devs experimenting. Not the AI-curious middle. The most senior, most experienced people, the ones who rarely wrote code day-to-day before LLMs, are now the heaviest users. That says something profound about where AI coding tools deliver the most leverage: not replacing junior work, but unblocking senior people who have deep architectural knowledge but limited time to implement.

The most practical takeaway for developers: invest in making your repositories agent-friendly before investing in fancier agents. Add pre-commit hooks, document environment variables, make builds self-verifiable. As @matanSF put it, fast validation loops make every agent more effective. The tooling is maturing fast, but it can only move as fast as your codebase lets it.

Quick Hits

  • @AnthropicAI published a new constitution for Claude, a detailed description of their vision for Claude's behavior and values, written primarily for Claude itself and used directly in training.
  • @scaling01 notes Anthropic is "preparing for the singularity," linking to what appears to be internal planning docs.
  • @mehulmpt declares "the end of ed-tech is near," a one-liner that hits different after Google launched free AI-powered SAT practice.
  • @Abhigyawangoo published "Why your AI agents still don't work," addressing the gap between agent hype and agent reality.
  • @__Talley__ made a Polymarket promo video in 30 minutes using only 4-5 prompts, adding to the growing pile of evidence that video editing workflows are being compressed dramatically.
  • @TheRealMcCoy shared a breakdown of photonic computing for AI, where light-based matrix multiplication could deliver massive speed gains with lower energy use. Still early but worth watching.
  • @theo pointed to the Claude Code ecosystem as the model for what good devrel looks like in 2026.
  • @dweekly shared that a Fortune 100 company liked to declare itself on the "frontier of AI" when only 1% of employees had access to any form of it.

Claude Code: From Tool to Ecosystem

Today felt like a tipping point for Claude Code's surrounding ecosystem. The sheer volume of community-built tooling, guides, and workflow patterns suggests the product has crossed from "powerful CLI" into "platform with a developer community."

The headline launch was Agentation from @benjitaylor, a visual feedback tool that lets you click elements, add notes, and copy markdown that gives agents element paths, selectors, and positions. It's the kind of tool that solves a real friction point: agents struggling to understand visual layouts. @benjitaylor went further, noting he "was able to build the entire documentation site solely using Claude Code + Agentation, including all the animated demos." A tool for agents, built by agents.

@affaanmustafa dropped "The Longform Guide to Everything Claude Code," covering token optimization, memory persistence, verification loops, parallelization, and subagent orchestration. It hit 7,500 stars and 1,000 forks in under four days. That kind of traction for a guide, not a tool, signals real hunger for operational knowledge about how to use these systems effectively.

On the tips-and-tricks front, @alexhillman shared several battle-tested patterns. The most immediately useful: injecting date-time context via hooks. "Claude Code doesn't know what time it is. Or what time zone you are in. So when you do date time operations of ANY kind, things get weird fast." His solution is a pre-message hook that generates current datetime and timezone, injected silently into context. Three months battle-tested, and it works. He also pushed a broader philosophy: "If you ask your AI assistant more questions than it asks you, you're gonna have a bad time." His approach combines confidence scoring with interviewing workflows, where agents stop and interview you when they're below a confidence threshold. @rezzz expanded on this, describing how he had the AI interview him about his fears, preferences, and working style so "the system works for me and not the other way around."

The Anthropic team shared their own war stories. @trq212 revealed they found a garbage collection issue in Claude Code's rendering pipeline that only surfaced in certain terminal/OS combinations, noting "some things you can't find until you ship." The underlying migration was massive: porting their entire rendering engine while keeping nothing user-facing broken, work that "could have taken on the order of 1-2 years for a single engineer" without Claude Code.

Other ecosystem moves: @simplifyinAI highlighted a new open-source library with 100+ pre-made agents, skills, and templates for Claude Code. @jakubkrcmar called out how @clawdbot is "quickly becoming the wet dream of leading AI companies." @jarredsumner announced Bun will support --cpu-prof-md, printing CPU profiles as Markdown so LLMs can read and grep them. Even runtime tooling is adapting to agent workflows. And @paraddox shared the simplest possible autonomous loop: a bash script running Claude 50 times with --dangerously-skip-permissions.

Agent Readiness and the Code Review Problem

Factory AI introduced Agent Readiness, a framework that measures how well a repository supports autonomous development across eight axes and five maturity levels. The framing is important: it's not about making better agents, it's about making better environments for agents to work in. @EnoReyes called it "the most essential focus area for a software organization looking to accelerate," warning that without it, "your adoption of AI will actively decelerate your org." @bentossell kept it simple: "all repos should be agent-ready."

@matanSF made the case concrete with examples that will resonate with anyone who's watched an agent flail: "No pre-commit hooks = agent waits 10 min for CI instead of 5 sec. Undocumented env vars = agent guesses, fails, guesses again. Build requires tribal knowledge from Slack = agent can't verify its own work." The pattern is clear. Agent effectiveness is bounded by repository quality.

The code review conversation ran in parallel. @ScottWu46 from Devin argued that current AI review tools focus on catching bugs at arm's length, but "until we reach the point where you can confidently hit 'Merge' on a 5000-line agent PR, you're still bottlenecked on reviewing the code yourself." The real question: would you rather have an AI that catches 80% of bugs, or an AI-powered review UX that makes you 5x faster? @walden_yan echoed this, saying "it felt pretty slop to say AI will review the code that it wrote. The key is going to be helping the HUMAN understand what they're merging." Meanwhile, @steveruizok offered a delightfully unorthodox approach: "The best code review tool I've come up with is asking Claude to reimplement the PR on a new branch in a narratively optimized perfect git history." @pauldix tied it together, arguing that "getting agents into a verification loop is the superpower for 2026."

Skills Discovery and the Context Layer

A potentially significant standards conversation emerged around how agents find and consume skills. @elithrar from Cloudflare proposed using the .well-known URI standard with an index of associated files, so agents can hit /.well-known/skills/index.json to discover related skills. "There's tools like add-skill to add skills, but you have to find them first." He acknowledged the tension around premature standardization but emphasized this is an RFC, not a standard, "big on the 'C' here."

Prefect went bigger. @jlowin announced Prefect Horizon, positioning it as "the context layer" where AI agents interface with your business. Built on top of FastMCP (which they created), Horizon adds managed hosting, a central registry, role-based access control, and an agentic interface for business users. The pitch: MCP tells you how to build a server, but not how to govern it at scale. Horizon aims to fill that gap. Whether the market needs an MCP platform layer this early is an open question, but the bet is clear. @LLMJunky demonstrated the power of the skills approach by generating a full promo video for CodexSkills from a single prompt, calling the result "cracked."

Products and Launches

Several product launches landed today. @tomkrcha launched Pencil, an infinite WebGL design canvas for Claude Code with parallel design agents and a git-native .pen file format. @theworldlabs opened the World API for building. @usekernel introduced Browser Pools with pre-configured logins, cookies, and extensions for agents, and @rfgarcia outlined use cases: spinning up browsers for QA, running evals on browser agents, and giving parallel subagents different research tasks without paying for standby CPU time.

Google entered the AI-for-education space by launching full-length, on-demand SAT practice exams in Gemini, grounded in content from Princeton Review, available at no cost. @Zai_org highlighted GLM Coding Plans paired with Kilo Code, focusing on the practical question of how much work you can get done without worrying about limits or cost.

AI and the Career Landscape

@hamptonism posted the meme of the day: "POV: driving to your $450k SWE job knowing it's just another 8 hours of having Claude do everything for you until you're eventually replaced entirely within 12 months." It's a joke, but @GergelyOrosz's observation about distinguished engineers and VPs dominating internal token leaderboards adds a serious dimension. The most experienced people are getting the most leverage, which suggests AI tools amplify expertise rather than replacing it.

@esrtweet invoked Vernor Vinge's Singularity concept, arguing we're living in it right now: "Nobody knows what to build that will still have value in 3 months." The practical implication for planning is real. Product roadmaps built on 12-month horizons are struggling when the underlying capabilities shift quarterly.

Local AI: Big Models on Small Hardware

@LiorOnAI highlighted AirLLM, which runs 70B parameter models on 4GB VRAM by loading one layer at a time, computing, freeing memory, and loading the next. It can reportedly run Llama 3.1 405B on 8GB VRAM. No quantization required by default, same API as Hugging Face Transformers. The tradeoff is speed for memory, but for prototyping and testing, the accessibility is compelling.

@DataChaz covered NVIDIA's PersonaPlex-7B, an open-source full-duplex conversational model released under MIT license. Unlike traditional ASR-to-LLM-to-TTS pipelines that force rigid turn-taking, PersonaPlex uses a dual-stream transformer to listen and speak simultaneously, enabling "instant back-channel responses, interruptions that feel human, real conversational rhythm." For anyone building voice agents or low-latency assistants, this is a meaningful step toward natural conversation.

Source Posts

๐Ÿ“™
๐Ÿ“™ Alex Hillman @alexhillman ·
@stolinski Add a user message hook that uses bash to check the date and time. Injects it into the session invisible to you but reminds the agent what time it is.
G
Google @Google ·
Weโ€™re launching full-length, on demand practice exams for standardized tests in @GeminiApp, starting with the SAT, available now at no cost. Practice SATs are grounded in rigorously vetted content in partnership with @ThePrincetonRev, and Gemini will provide immediate feedback highlighting where you excelled and where you might need to study more. To try it out, tell Gemini, โ€œI want to take a practice SAT test.โ€
W
World Labs @theworldlabs ·
Start building with the World API today: https://t.co/vGQ0MfDfYT
J
Jarred Sumner @jarredsumner ·
In the next version of Bun `bun --cpu-prof-md <script>` prints a CPU profile as Markdown so LLMs like Claude can easily read & grep it https://t.co/1B3Xv3pcLG
W
Walden @walden_yan ·
What do we actually need to review code 10x faster? It felt pretty slop to say AI will review the code that it wrote. The key is going to be helping the HUMAN understand what theyโ€™re merging. So we built a new interface for this
C Cognition @cognition

Meet Devin Review: a reimagined interface for understanding complex PRs. Code review tools today donโ€™t actually make it easier to read code. Devin Review builds your comprehension and helps you stop slop. Try without an account: https://t.co/Zzu1a3gfKF More below ๐Ÿ‘‡ https://t.co/sYQLjwSk6s

L
Lior Alexander @LiorOnAI ·
You can now run 70B LLMs on a 4GB GPU. AirLLM just made massive models usable on low-memory hardware. ๐—ช๐—ต๐—ฎ๐˜ ๐—ท๐˜‚๐˜€๐˜ ๐—ต๐—ฎ๐—ฝ๐—ฝ๐—ฒ๐—ป๐—ฒ๐—ฑ AirLLM released memory-optimized inference for large language models. It runs 70B models on 4GB VRAM. It can even run 405B Llama 3.1 on 8GB VRAM. ๐—›๐—ผ๐˜„ ๐—ถ๐˜ ๐˜„๐—ผ๐—ฟ๐—ธ๐˜€ AirLLM loads models one layer at a time. Instead of loading everything: โ†’ Load a layer โ†’ Run computation โ†’ Free memory โ†’ Load the next layer This keeps GPU memory usage extremely low. ๐—ž๐—ฒ๐˜† ๐—ฑ๐—ฒ๐˜๐—ฎ๐—ถ๐—น๐˜€ โ€ข No quantization required by default โ€ข Optional 4-bit or 8-bit weight compression โ€ข Same API as Hugging Face Transformers โ€ข Supports CPU and GPU inference โ€ข Works on Linux and macOS Apple Silicon ๐—ช๐—ต๐—ฎ๐˜ ๐˜†๐—ผ๐˜‚ ๐—ฐ๐—ฎ๐—ป ๐—ฑ๐—ผ โ€ข Run Llama, Qwen, Mistral, Mixtral locally โ€ข Test large models without cloud GPUs โ€ข Prototype agents on cheap hardware
K
Kevin @kcosr ·
@FactoryAI Who is going to create an open skill for this concept? Any takers? ๐Ÿค”
E
Eric S. Raymond @esrtweet ·
We're in the Singularity now, and it's screwing up the business planning of everybody in tech. How do you do product design when the pace of change in AI is so rapid that you can be pretty sure your concept will be obsolete before it ships? Vernor Vinge first articulated the concept of the Singularity in 1983, describing it as the point at which technological change accelerates to a speed where what comes after the Singularity is incomprehensible in terms of what was before it. And that's right where we are in early 2026. Nobody knows what to build that will still have value in 3 months. Which, in retrospect...what did you think it was going to be like? Vibes? Papers? Essays? Strap in, kids. The ride is only going to get wilder.
B Ben @bwarrn

Lunch w/ an exited founder who helps fortune 500 companies adopt AI. Insane reality check: Some of the biggest companies on earth use *zero* AI tools. Not even ChatGPT. Execs only recognize: ChatGPT, Copilot, Gemini (maybe Perplexity). Everyone feels behind. Nobody knows what to buy or how to plug it in. The "AI saturation" narrative is another example of what a bubble Silicon Valley is. Rest of the world hasnโ€™t started yet. We have to build for the 99%.

J
Jeremiah Lowin @jlowin ·
Every company will have a context layer. Today, we're launching Prefect Horizon to make that possible. The context layer is where AI agents interface with your business. It's where teams expose their proprietary data, tools, and workflows to autonomous systems. Itโ€™s where you define what agents can see and do, and where you curate and distribute the context that makes automation useful and safe. We know what this requires because we created FastMCP, the standard framework for working with the Model Context Protocol. We've watched it grow to a million downloads a day. But we've also seen the protocol's limits. MCP describes how to *build* a server, not how to govern it at scale. It doesn't tell you how to deploy it, secure it, or share it with 5,000 colleagues. We built Prefect Horizon to turn MCP from a protocol into a platform: ๐Ÿš€ Horizon Deploy: Managed MCP hosting that gets you from PR to governed endpoint in 60 seconds. ๐Ÿ“š Horizon Registry: A central catalog for all your MCP servers, including first-party, third-party, and our new remix servers that compose multiple sources together. ๐Ÿ” Horizon Gateway: The control plane enterprises need: role-based access control down to the tool level, audit logs, and usage visibility. ๐Ÿ’ฌ Horizon Agents: The last mile. Give business users an agentic interface to your company without ever knowing what "MCP" is. This is automation for the context era. Welcome to Horizon.
๐Ÿ“™
๐Ÿ“™ Alex Hillman @alexhillman ·
hillman's razer of ai assistants: if you ask your AI assistant more questions than it asks you, you're gonna have a bad time. the real magic is combining confidence scoring with interviewing workflows. effectively "if you're not above X confidence threshold, stop and use this interview workflow until you're above that threshold" solves a wide swath of problems
J Jason Resnick ๐ŸŒฒ๐Ÿ’Œ @rezzz

@theirongolddev @alexhillman What Alex did I thought was geniusโ€ฆ I had it interview me for ergonomics I had it ask me my fears, what I didnโ€™t like, what works for me, what I want, how I want to work/show up, and other things about me so the system works for me and not the other way around.

A
Anthropic @AnthropicAI ·
Weโ€™re publishing a new constitution for Claude. The constitution is a detailed description of our vision for Claudeโ€™s behavior and values. Itโ€™s written primarily for Claude, and used directly in our training process. https://t.co/CJsMIO0uej
T
Tom Osman ๐Ÿฆโ€โฌ› @tomosman ·
How I'm using Clawd.bot to change how I get things done.
L
Lior Alexander @LiorOnAI ·
Repo: https://t.co/NAqzYcIyua
C
Chris McCoy @TheRealMcCoy ·
Fascinating. tl;dr for my crowd Photonic computing swaps electricity for light to handle the massive number-crunching that makes AI models work, particularly the matrix multiplications needed to train and run large systems like ChatGPT. Light travels extremely fast and can process huge amounts of data all at once through beams spreading out, overlapping, or using different colors (wavelengths), hitting speeds around 100 trillion cycles per second. Recent breakthroughs in top scientific journals show setups where these giant multiplications happen in a single quick pass of lightโ€”meaning the time it takes doesn't grow much bigger even when dealing with enormous models or datasets, unlike regular computer chips that slow down as things get larger. This could bring huge jumps in speed and much lower energy use for AI tasks, potentially shifting future computers to rely mainly on light instead of electrical signals.
T
Thariq @trq212 ·
This was a legacy migration, we had to port our entire rendering engine while making sure nothing user-facing broke. Doing this without Claude Code could have taken on the order of 1-2 years for a single engineer, something we would have never been able to prioritize.
a
am.will @LLMJunky ·
Holy sh*t! This is cracked. I just ran this skill in my repo with the following prompt: 'Make me a flash promo video for CodexSkills that shows installing the skills and then highlights all the skills available.' And it came up with this without any further prompting. ๐Ÿคฏ Are you kidding me?
R Remotion @Remotion

Remotion now has Agent Skills - make videos just with Claude Code! $ npx skills add remotion-dev/skills This animation was created just by prompting ๐Ÿ‘‡ https://t.co/hadnkHlG6E

T
Theo - t3.gg @theo ·
If youโ€™re wondering what good โ€œdevrelโ€ looks like in 2026, itโ€™s this
W Wayne Sutton @waynesutton

Now you can track your @opencode and @claudeai CLI coding sessions in one place. https://t.co/FLe8dRC8Pv provides searchable history, markdown export, and eval-ready datasets. See tool usage, token spend, and session activity across projects. Check out the demo. https://t.co/HGlZOOyugN

B
Benji Taylor @benjitaylor ·
I was able to build the entire documentation site solely using Claude Code + Agentation, including all the animated demos. Check out the full docs here: https://t.co/FRyZMEQn5Y
๐Ÿ“™
๐Ÿ“™ Alex Hillman @alexhillman ·
I meet a lot of people who don't realize how much valuable paper trail Claude Code creates for itself. Slurping up those session transcripts and parsing them in various ways unlocks: - memory and recall - pattern recognition - self-generating/repairing skills and workflows And SO MUCH MORE
T Thariq @trq212

@souravbhar871 Itโ€™s all stored locally in your .claude folder, you can ask Claude to read it and create scripts to help visualize it

M
Mehul Mohan @mehulmpt ·
The end of ed-tech is near
G Google @Google

Weโ€™re launching full-length, on demand practice exams for standardized tests in @GeminiApp, starting with the SAT, available now at no cost. Practice SATs are grounded in rigorously vetted content in partnership with @ThePrincetonRev, and Gemini will provide immediate feedback highlighting where you excelled and where you might need to study more. To try it out, tell Gemini, โ€œI want to take a practice SAT test.โ€

P
Paul Dix @pauldix ·
Getting agents into a verification loop is the superpower for 2026. Agents will build all the software if you give them the context and tools to verify and iterate. My thoughts on Building the Machine that Builds the Machine:
P Paul Dix @pauldix

Build the machine that builds the machine

a
abhi @Abhigyawangoo ·
Why your AI agents still donโ€™t work
L
Lisan al Gaib @scaling01 ·
Anthropic is preparing for the singularity https://t.co/QtTehqoyu8
L Lisan al Gaib @scaling01

I'm starting to get worried. Did Anthropic solve continual learning? Is that the preparation for evolving agents? https://t.co/pcCoSM4gAr

K
Kernel @usekernel ·
Introducing Browser Pools โ€” instant browsers with the logins, cookies, and extensions your agents depend on. Designed to make using Kernel even faster. https://t.co/Gt6cc9awcd
โ‚•
โ‚•โ‚โ‚˜โ‚šโ‚œโ‚’โ‚™ @hamptonism ·
pov: driving to your $450k swe job knowing itโ€™s just another 8 hours of having Claude do everything for you until youโ€™re eventually replaced entirely within 12 months, https://t.co/AclKNRZCKP
M
Matt Silverlock ๐Ÿ€ @elithrar ·
instead of trying to find them in a repo (or is it a second skills/* repo?), in docs, or <elsewhere>: the idea is that you agents can hit /.well-known/skills/index.json to discover related skills.
c
cogsec @affaanmustafa ·
The Longform Guide to Everything Claude Code
S
Scott Wu @ScottWu46 ·
Most AI review tools today center around asking an arms-length agent to catch & report potential bugs. This is really valuable! But until we reach the point where you can confidently hit "Merge" on a 5000-line agent PR, you're still bottlenecked on reviewing the code yourself. This will stay true for a while even as the tools get better. Would you rather have an arms-length AI that catches 80% of bugs or an AI-powered review UX that makes *you* 5x faster? Probably the latter since you'd still have to review the whole PR yourself to catch the last 20%. Of course, the best review experience should have both! We built Devin Review with these thoughts in mind. Let us know what you think!
C Cognition @cognition

Meet Devin Review: a reimagined interface for understanding complex PRs. Code review tools today donโ€™t actually make it easier to read code. Devin Review builds your comprehension and helps you stop slop. Try without an account: https://t.co/Zzu1a3gfKF More below ๐Ÿ‘‡ https://t.co/sYQLjwSk6s

E
Eno Reyes @EnoReyes ·
Agent Readiness is the most essential focus area for a software organization looking to accelerate. As an engineering leader itโ€™s your responsibility to start this effort now. Without it, your adoption of AI will actively decelerate your org. Very important to get right!
F Factory @FactoryAI

Introducing Agent Readiness. AI coding agents are only as effective as the environment in which they operate. Agent Readiness is a framework to measure how well a repository supports autonomous development. Scores across eight axes place each repo at one of five maturity levels. https://t.co/9POPIY3hXr

S
Simplifying AI @simplifyinAI ·
Claude Code just got an "App Store" for agents ๐Ÿคฏ A massive new open-source library just dropped with 100+ pre-made agents, skills, and templates that you can install instantly. And it's 100% free to use. https://t.co/2R07ziHNKV
S
Steve Ruiz @steveruizok ·
The best code review tool I've come up with is asking Claude to reimplement the PR on a new branch in a narratively optimized perfect git history
S Steve Ruiz @steveruizok

v1 of my "reimplement this PR using an ideal commit history" command, actually works quite well. "What commits would I have made if I had perfect information about the desired end state?" https://t.co/5S4kCIo8bR

F
Factory @FactoryAI ·
Introducing Agent Readiness. AI coding agents are only as effective as the environment in which they operate. Agent Readiness is a framework to measure how well a repository supports autonomous development. Scores across eight axes place each repo at one of five maturity levels. https://t.co/9POPIY3hXr
J
Jason Resnick ๐ŸŒฒ๐Ÿ’Œ @rezzz ·
@theirongolddev @alexhillman What Alex did I thought was geniusโ€ฆ I had it interview me for ergonomics I had it ask me my fears, what I didnโ€™t like, what works for me, what I want, how I want to work/show up, and other things about me so the system works for me and not the other way around.
M
Matt Silverlock ๐Ÿ€ @elithrar ·
We're proposing an approach to make discovering agent skills easier by using the .well-known URI standard + a index of associated files for agents to pull down. There's tools (like add-skill) to add skills, but you have to find them first. RFC: https://t.co/9Qrf0aXJ5D
c
cogsec @affaanmustafa ·
~7500 stars and ~1000 forks in < 4 days 01/21/2026 @ 9AM PST: "The Longform Guide to Everything Claude Code" > Token optimization > Memory persistence > Continuous Learning > Verification loops > Parallelization > Subagent orchestration > + advanced e.g. (pass@k vs pass^k) https://t.co/0pvpQDc5CP
c cogsec @affaanmustafa

The Shorthand Guide to Everything Claude Code

B
Benji Taylor @benjitaylor ·
Introducing Agentation: a visual feedback tool for agents. Available now: ~npm i agentation Click elements, add notes, copy markdown. Your agent gets element paths, selectors, positions, and everything else it needs to find and fix things. Link to full docs below โ†“ https://t.co/o65U5MY7V6
M
Matt Silverlock ๐Ÿ€ @elithrar ·
@mitsuhiko to be clear: I think we (as an industry) standardize things WAY too early. however: I don't consider an RFC (certainly a non-IETF one like this) a standard. big on the "C" here!
T
Tom Krcha @tomkrcha ·
Excited to launch Pencil INFINITE DESIGN CANVAS for Claude Code > Superfast WebGL canvas, fully editable, running parallel design agents > Runs locally with Claude Code โ†’ turn designs into code > Design files live in your git repo โ†’ Open json-based .pen format https://t.co/UcnjtS99eF
D
Ddox @paraddox ·
You folks asked for it. Simplest Ralph loop: #!/bin/bash PROMPT="${1:-prompt here}" for i in {1..50}; do echo "=== Run $i/50 ===" claude --dangerously-skip-permissions -p "$PROMPT" echo "" done
Z
Z.ai @Zai_org ·
Amazing blog from @kilocode ๐Ÿ‘‡ "The real question isnโ€™t 'whatโ€™s the smartest model?' Itโ€™s 'how much real work can I get done without constantly worrying about limits or cost?' Thatโ€™s the gap GLM Coding Plans are meant to fill, especially when paired with Kilo Code." https://t.co/7C6oqCnNkD
D
David E. Weekly @dweekly ·
@bwarrn I worked for a Fortune 100 company that liked to declare itself on the "frontier of AI" when only one percent of the employee population had access to any form of it.
C
Charly Wargnier @DataChaz ·
NVIDIA just removed one of the biggest friction points in Voice AI. PersonaPlex-7B is an open-source, full-duplex conversational model. Free, open source (MIT), with open model weights on @huggingface ๐Ÿค— Links to repo and weights in ๐Ÿงตโ†“ The traditional ASR โ†’ LLM โ†’ TTS pipeline forces rigid turn-taking. Itโ€™s efficient, but it never feels natural. PersonaPlex-7B changes that. This @nvidia model can listen and speak at the same time. It runs directly on continuous audio tokens with a dual-stream transformer, generating text and audio in parallel instead of passing control between components. That unlocks: โ†’ instant back-channel responses โ†’ interruptions that feel human โ†’ real conversational rhythm Persona control is fully zero-shot! If youโ€™re building low-latency assistants or support agents, this is a big step forward ๐Ÿ”ฅ
G
Gergely Orosz @GergelyOrosz ·
One interesting observation: inside a Big Tech, the internal token leaderboard is dominated byโ€ฆ very very experienced engineers. Distinguished-level folks who you rarely saw code day to day before LLMs. Also, some VPs (!!)
B
Ben Tossell @bentossell ·
all repos should be agent-ready
F Factory @FactoryAI

Introducing Agent Readiness. AI coding agents are only as effective as the environment in which they operate. Agent Readiness is a framework to measure how well a repository supports autonomous development. Scores across eight axes place each repo at one of five maturity levels. https://t.co/9POPIY3hXr

T
Talley @__Talley__ ·
Okayโ€ฆ video editors are cooked. I made this video for Polymarket in 30 minutes. Only took 4-5 prompts. https://t.co/YFOeHSTwgW
R Remotion @Remotion

Remotion now has Agent Skills - make videos just with Claude Code! $ npx skills add remotion-dev/skills This animation was created just by prompting ๐Ÿ‘‡ https://t.co/hadnkHlG6E

R
Rafael Garcia @rfgarcia ·
Browser pools unlock so many cool uses cases: - Spin up a of bunch of browsers all QAing your site - Run large-scale evals on your browser agent - Give a fleet of parallel subagents different research tasks Keep them running as long as you like w/o getting charged for standby CPU time.
K Kernel @usekernel

Introducing Browser Pools โ€” instant browsers with the logins, cookies, and extensions your agents depend on. Designed to make using Kernel even faster. https://t.co/Gt6cc9awcd

J
Jakub Krcmar @jakubkrcmar ·
Its nuts to see what an open source project like @clawdbot is quickly becoming โ€” wet dream of leading ai companies and many startups. Just shows how fundamental things are shifting. Respect to @steipete
N Nat Eliason @nateliason

Yeah this was 1,000% worth it. Separate Claude subscription + Clawd, managing Claude Code / Codex sessions I can kick off anywhere, autonomously running tests on my app and capturing errors through a sentry webhook then resolving them and opening PRs... The future is here.

T
Thariq @trq212 ·
While we test Claude Code rigorously, our users run Claude in a huge combination of terminal and operating system setups. Here we found that in some setups we were triggering Garbage Collection too often in our rendering pipeline. Somethings you can't find until you ship.
M
Matan Grinberg @matanSF ·
โ€ข No pre-commit hooks = agent waits 10 min for CI instead of 5 sec โ€ข Undocumented env vars = agent guesses, fails, guesses again โ€ข Build requires tribal knowledge from Slack = agent can't verify its own work codebases with fast validation makes every agent more effective
F Factory @FactoryAI

Introducing Agent Readiness. AI coding agents are only as effective as the environment in which they operate. Agent Readiness is a framework to measure how well a repository supports autonomous development. Scores across eight axes place each repo at one of five maturity levels. https://t.co/9POPIY3hXr

M
Matt Silverlock ๐Ÿ€ @elithrar ·
e.g. https://t.co/QVwuPE2YHO has skills for the agents-sdk, durable-objects, wrangler and cloudflare (just to start).
๐Ÿ“™
๐Ÿ“™ Alex Hillman @alexhillman ·
When I started building my assistant I figured this one out FAST. Claude Code doesn't know what time it is. Or what time zone you are in. So when you do date time operations of ANY kind, as simple as saving something to your calendar, things get weird fast. My early solution has stuck thru every iteration of my JFDI system and it's dummy simple: I use Claude Code hooks to run a bash script that generates current date time, timezone of host device, friendly day of week etc. Injects it silently into context. I never see it but date time issues vanish. 3+ most battle tested. Kinda wild that this isn't baked in @bcherny (thank you for CC btw it changed my life no exaggerating)
S Scott Tolinski - Syntax.fm @stolinski

My clawdbot sucks at days and time. It never seems to have any clue what the current day or time is.