AI Digest.

The Claude Code Playbook Crystallizes as Cowork Launches and Node.js Ships Critical Security Fix

The developer community converged on Claude Code best practices with multiple viral threads on CLAUDE.md configurations, TDD workflows, and agent coding patterns. Anthropic's Claude Cowork launch prompted one startup to open-source their competing product overnight. A critical Node.js security vulnerability affecting virtually every production app demanded immediate patching.

Daily Wrap-Up

January 13th was the day agent-assisted coding stopped being experimental and started growing a canon. Multiple threads on Claude Code best practices hit simultaneously, each approaching the same problem from a different angle: how do you actually get reliable, repeatable results from AI coding agents? The answers converging from @ericzakariasson, @mattpocockuk, @alexhillman, and others aren't revolutionary individually, but taken together they paint a picture of a discipline forming in real time. TDD-first workflows, carefully structured CLAUDE.md files, rules versus skills distinctions. These aren't tips anymore. They're becoming the standard operating procedure.

The other major current running through today's posts is a collective realization about agent capability. @davis7 captured it perfectly, admitting they had deliberately avoided pushing agents hard because the implications of them being truly capable felt threatening. @levie framed it more clinically: there's a massive capability overhang right now, with most organizations still treating AI as a chatbot rather than an autonomous worker. Both perspectives point to the same conclusion. The gap between what agents can do and what most people use them for is widening, not narrowing. The teams that close that gap first will have a serious advantage.

On the security front, Node.js shipped patches for a critical vulnerability that hits React Server Components, Next.js, and every major APM tool. Not glamorous, but the kind of thing that separates teams that keep their infrastructure tight from those that find out the hard way. The most practical takeaway for developers: if you're using Claude Code or any AI coding agent, start building your CLAUDE.md and rules files today. The emerging consensus is clear: agents perform best when given explicit, structured context about your codebase, your patterns, and your expectations. Think of it as documentation that actually pays dividends every session.

Quick Hits

  • @kr0der almost quit Codex after one day but found the right workflow. The tool gap is often a knowledge gap.
  • @dabit3 explored Claude's new programmatic tool calling beta, which runs code in a sandbox before returning results to the model, cutting latency and token consumption.
  • @TheAhmadOsman posted an exhaustive project-based LLM engineering curriculum covering everything from tokenization to MoE routing. Ambitious, but the "build, plot, break, repeat" philosophy is sound.
  • @emollick suggested organizations should describe what they do in plain English markdown files. Quietly one of the most impactful pieces of advice for anyone adopting AI tooling.
  • @johnrushx dropped startup wisdom they wish they'd had earlier.
  • @clawdbot shipped v2026.1.12 with vector memory and voice call capabilities.
  • @pepicrft released a Clawdbot Vault Plugin that turns a local folder into a structured knowledge vault with QMD-powered search and embeddings.
  • @hive_echo shared nano banana pro UI mockups.
  • @tyler_agg published a guide on making realistic longform AI videos with prompts included.
  • @Oxylabs_io promoted an all-in-one Web Scraper API with task scheduling and crawling.

The Claude Code Playbook Takes Shape

The sheer volume of Claude Code content today suggests the community has hit an inflection point where early adopters are codifying their hard-won patterns into shareable knowledge. What started as individual experimentation is becoming collective wisdom, with at least eight separate posts contributing different pieces of the same puzzle: how to make AI coding agents consistently useful rather than intermittently impressive.

@ashpreetbedi shared their personal Claude Code workflow, while @rohit4verse surfaced how the creator of Claude Code actually writes software. @twannl, who spends most of their time in Cursor, called out one particular article as a must-read, noting they learned a lot despite already being deep in agent-assisted development. @Hesamation pointed to what they consider the definitive Claude Code guide:

> "this is still the best guide on Claude Code I've seen that covers basically how you should (and shouldn't) use it. comprehensive, practical, and to-the-point." — @Hesamation

The CLAUDE.md file emerged as a recurring theme. @mattpocockuk shared specific additions that made plan mode dramatically more useful, turning "unreadably long plans" into "concise, useful plans with followup questions." @alexhillman took a different angle, focusing on communication style and anti-patterns. Their additions read like a style guide for AI interaction, banning ellipses ("comes across as passive aggressive"), hedging phrases, and what they categorized as "AI Slop Patterns":

> "Never use 'not X, but Y' or 'not just X, but Y' - state things directly. No hedging: 'I'd be happy to...', 'I'd love to...', 'Let me go ahead and...' No performative narration: Don't announce actions then do them - just do them" — @alexhillman

@aye_aye_kaplan shared the Cursor team's official recommendations for coding with agents, acknowledging how rapidly best practices are evolving. The pattern across all these posts is unmistakable: the tooling has matured enough that the bottleneck has shifted from capability to configuration. The developers getting the most value aren't necessarily the most skilled programmers. They're the ones who've invested in setting up their environment properly.

Agent Coding Patterns: TDD, Rules, and Encoded Expertise

Beyond general Claude Code usage, a more specific conversation emerged about concrete coding patterns that work well with agents. @ericzakariasson posted a detailed thread distilling what separates effective agent users from frustrated ones, with TDD emerging as a particularly powerful pattern:

> "have agent write tests (explicit TDD, no mock implementations) - run tests, confirm they fail - commit tests - have agent implement until tests pass - commit implementation. agents perform best when they have a clear target to iterate against" — @ericzakariasson

The thread also drew a useful distinction between rules and skills. Rules are static context loaded into every conversation (code style, commands, workflow instructions), while skills are dynamic capabilities loaded when relevant. The advice to "start simple, add rules only when you see repeated mistakes" echoes the broader principle that emerged today: configuration should be driven by observed problems, not theoretical completeness.

@rauchg announced that Vercel is taking this idea to its logical conclusion, encoding over ten years of React and Next.js frontend optimization knowledge into reusable agent skills, distilled from engineers like @shuding. This is a significant move. Rather than hoping developers read documentation, Vercel is packaging expertise directly into the agent workflow. @PrajwalTomar_ demonstrated the practical impact, building a landing page with scrollytelling animations in under ten minutes using Cursor and Opus 4.5.

The synthesis here matters: TDD gives agents verifiable goals, rules give them consistent context, and encoded skills give them domain expertise. Stack all three and you get something that starts to look less like autocomplete and more like a junior developer who actually reads the docs.

The Agent Capability Inflection Point

Several posts today captured a moment of collective reckoning about what AI agents can actually do when pushed. @davis7 wrote the most candid version of this realization, admitting they had deliberately avoided testing agent capabilities because the implications were uncomfortable:

> "I very deliberately believed that agents weren't capable of anything 'real' because I honestly didn't want them to be. It was so much easier to just think it's not possible to do the very real and serious and important real engineering things I do" — @davis7

@levie provided the strategic framing, arguing that a "massive capability overhang" exists because most organizations still think of AI as chatbots. The winners in 2026, per @levie, will be those who figure out the right agent scaffolding, the right context engineering, and the change management to actually shift workflows. @blader was more blunt: "every company should be rolling their own devin like ramp. It will take you less than a day to standup and maybe a week to make good."

@marcelpociot illustrated what this looks like in practice, describing how Cowork was shipped in just a week and a half with human developers meeting in person for architectural decisions while each managing three to eight Claude instances simultaneously. @io_sammt predicted a new class of technician born in 2026, capable of building complex production systems in minutes. Whether that timeline is optimistic or not, the direction is clear: the gap between knowing agents are capable and actually leveraging that capability is where the value lives right now.

Multi-Agent Tools: Cowork, AgentCraft, and ralph-tui

The tooling around multi-agent orchestration had a busy day. @dejavucoder introduced Claude Cowork, Anthropic's entry into collaborative agent workflows. The launch had immediate competitive ripple effects. @guohao_li posted with admirable candor:

> "Anthropic Claude Cowork just killed our startup product. So we did the most rational thing: open-sourced it. Meet Eigent" — @guohao_li

The open-sourcing of Eigent is a pattern we've seen before: when a platform vendor ships a feature that competes with your product, the fastest path to relevance is giving away the code and building community around it. Whether Eigent gains traction remains to be seen, but the move itself signals how quickly the multi-agent space is consolidating around major providers.

On the indie side, @theplgeek shipped ralph-tui, a terminal UI for managing agentic coding loops with PRD creation and task management built in. @idosal1 took a more playful approach with AgentCraft, letting developers orchestrate agents through an RTS game interface. These tools represent a growing recognition that managing multiple agents requires dedicated interfaces, not just more terminal tabs. The orchestration layer is becoming its own product category.

Node.js Critical Security Patch and Open-Source PII Detection

Not everything today was about agents. @matteocollina flagged a critical Node.js security release affecting "virtually every production Node.js app," specifically calling out React Server Components, Next.js, and APM tools like Datadog, New Relic, and OpenTelemetry as vulnerable to DoS attacks. The official @nodejs account confirmed patches across four release lines (25.x, 24.x, 22.x, 20.x) addressing three high-severity, four medium-severity, and one low-severity issue. If you're running Node.js in production and haven't patched yet, stop reading and go update.

In a separate but thematically related development, @MaziyarPanahi highlighted OpenMed's mass release of 35 PII detection models under Apache 2.0 licensing, specifically targeting healthcare AI safety with HIPAA and GDPR compliance. The open-source release of production-grade safety tooling is a welcome counterbalance to the speed-first culture that dominates most AI development discourse. Building fast is great. Building fast without leaking patient data is better.

Sources

A
Addy Osmani @addyosmani ·
AI may change how we do code reviews. PRs show what changed. Prompt logs show what the human actually wanted. Full trajectories - the conversation, the iterations, the steering - show you how they got there. When agents write the code, review inverts. You stop asking only "is this correct?" and start asking "was this intent clear enough to execute safely?" Most teams won't abandon code review. They'll do both. Review the output for correctness, review the trajectory for intent. The diff tells you what shipped. The conversation tells you why. We're not replacing PRs but we may consider the prompt is the spec, the code is the build output, and review should also happen at the layer where human judgment actually lives.
G GergelyOrosz @GergelyOrosz

"I don't like pull requests (PRs) any more. A large chunk code change doesn't tell me much about the intent or why it was done. I now prefer prompt requests. Just share the prompt you ran / want to run. If I think it's good, I'll run it myself and merge it." - @steipete wow

B
Bilgin Ibryam @bibryam ·
"The best software engineers won’t be the fastest coders, but those who know when to distrust AI." The Next Two Years of Software Engineering - @addyosmani https://t.co/gcR3b75Mpu
A
Angry Tom @AngryTomtweets ·
@antoinemarcel this is Kling AI 2.6 Motion Control
J
Jaime Jorge @jaimefjorge ·
The biggest takeaways/nuggets from my interview with @GeoffreyHuntley on AI-native software engineering and the Ralph loop: 1. Software development and software engineering are now two different professions, and one of them is over. Software development, the work of translating tickets into code, can now be done by anyone for $10-42/hour while they sleep. Software engineering, architecture, security, requirements breakdown, understanding failure modes, is where humans still matter. If you identify as a "software developer," you're competing against a bash loop. If you identify as a "software engineer," your job is to orchestrate the loops. 2. The moat you think protects your software product doesn't exist anymore. Geoffrey argues you can clone any SaaS product, even those with BSL licenses or proprietary enterprise code, using AI. He ran Ralph in reverse on HashiCorp Nomad's source code to generate clean-room specifications. When he hit gaps from missing enterprise features, he ran Ralph over their marketing materials and product docs to fill them in. Any company relying on licensing or code secrecy as a competitive moat needs to rethink their strategy. 3. Cursor, Windsurf, and every other AI coding tool are essentially the same thing: a loop that automatically copies and pastes. Geoffrey built these tools professionally and says the harness does almost nothing; the model does all the work. There's no real moat in the harness business when you're reselling tokens. The only differentiator is taste and UX. Stop evaluating tools and start learning the underlying patterns. 4. Ralph is not a product. It's an orchestrator pattern for running thousands of AI loops. The simplest version is a bash loop that deterministically allocates memory, lets the LLM pick one task, executes it, then starts fresh. The key insight: every loop gets a brand new context window. You avoid compaction (where the AI gets dumber as context fills up) by never letting the context window accumulate competing goals. Your institutional knowledge lives in specification files, not in the context window. 5. Specifications are the new source code. Geoffrey's workflow: spend 30 minutes in conversation with AI, drilling into requirements, making engineering decisions, building up specs. Then throw those specs to Ralph and get weeks worth of work in hours. The specs act as a "pin" that reframes every fresh loop with your domain knowledge. He doesn't hand-write specs. He code-generates them through structured conversation. Prototypes are now free. Refactoring is cheap. 6. The entry-level path into software engineering is closing fast. Geoffrey's company stopped hiring juniors for a year until they figured out how to interview for AI-native skills. There's already a cohort of juniors who've been practicing these techniques for six months. They'll work at a quarter of senior wages and outship them. If you're just picking up these tools today, you're behind. The new interview question: can you explain how to build a coding agent on a whiteboard? 7. Senior engineers who refuse to adapt are in more danger than juniors who embrace it. Geoffrey sees respected engineers taking hardline stances against AI ("it's installing fascism in your codebase"). Meanwhile, leadership teams are discovering Ralph and realizing three people can run the output of an entire org. When commit velocity and product velocity diverge that dramatically between adopters and non-adopters, founders notice. The hard line is coming. 8. AI is an amplifier of operator skill, not a replacement for it. If you're great at security and you get good at AI, you become a weapon. If you're mediocre and you use AI, you're still mediocre, just faster. The skill gap comes from "discoveries": learning the tricks, the loop-backs, the ways to close the automation loop. These techniques don't have standardized language yet. We're inventing the terms for the new computer every day. 9. Open source may no longer make sense for most use cases. Geoffrey, a former prominent open source maintainer whose land was funded by Open Collective, no longer uses open source libraries. His reasoning: every dependency injects a human into the loop. If there's a bug, you open a PR, chase a maintainer, wait. That's not automation. Instead, code-generate what you need. The exception: don't generate cryptography or security-critical code unless you have the domain expertise to verify it. 10. Programming languages now have a tier list based on how well AI agents can work with them. S-tier: Rust, TypeScript (especially with Effect.js), Python with Pydantic. These are source-based with strong type systems that reject invalid generations and work well with ripgrep for code discovery. F-tier: Java and .NET. Their DLL-based dependency systems don't work natively with the search tools AI agents use. The tradeoff with Rust: compilation is slow, so bad generations cost more time. 11. Corporate AI transformation programs are dangerously slow. Three-to-four-year rollouts with coaches and committees won't cut it when three founders in Bali can Ralph your entire product and undercut your pricing by 99%. Smaller teams ship faster. By the time the transformation is done, the market has moved. Geoffrey calls this the "Titanic moment": the boat is full, get the next boat. 12. We have a new computer, and that's why the legends are coming out of retirement. The last 40 years of computing decisions were designed for humans: TTYs, environment variables, slow language evolution to avoid breaking mental models. Now we have robots. What's the bare minimum a robot needs? Geoffrey sees this as the most exciting time in computing. If you're not excited about what you can now build, you haven't truly picked up the new computer yet.
V
Victor @victor_explore ·
@DanielGlejzner the real context window was the architecture decisions we made along the way
K
Klaas @forgebitz ·
having a monorepo turned out to be a massive advantage for ai coding all context is inside one repo api's, servers, auth, landing page, marketing sites, dashboard, ops, everything
H
Harry Charlesworth @hjcharlesworth ·
The gap is getting wider and I'm glad I could finally write this down. A mental model that works for us when pairing with an agent. https://t.co/xVFJG6JgM5
H
Harry Charlesworth @hjcharlesworth ·
Read it here: https://t.co/uniFTtCas6
📙
📙 Alex Hillman @alexhillman ·
I had my Claude assistant build a script to do them in batches. Local whisper model is free but slower. 200 would probably take a day or so. https://t.co/KFRjWr6VFf api keys work outbton $1-1.50/hr, but WAY faster, so for a few hundred bucks you can do the whole thing. My advice would be toget it to do one the way you want, THEN ask it to do a batch of 5 and see how it works/how much it costs, then ask it to do the full set
C
Codie Sanchez @Codie_Sanchez ·
Best money I've ever spent as a CEO... an internal AI transformation hire. He doesn't care about title. He just wants to ship. And he goes across your entire org, sales, revenue, hr, apps, tech and kills stupid manual processes. Such an underrated unlock.
M
Matthew Lam @mattlam_ ·
Fully set up my @clawdbot and now I have my 24/7 personal assistant + coding agent for $5/month. Easy to setup, I just got claude and codex to help me with Hetzner for VPS, and now I get some of my favorite use cases 24/7: - have a new project idea? Instead of just writing in my todo list, tell Clawdy (my assistant) to start helping me do relevant research, set up a new repo, or even start coding. - look through my task list, calendar, emails to help me plan my day and keep track of tasks - periodic reminders that I need (no longer need to go through Apple Reminders app just tell Clawdy) - X's search, including posts you've seen, I find pretty bad, I just get Clawdy to look for me with bird cli, much more likely to find a tweet I forgot to bookmark. @nikitabier checkout @steipete 's https://t.co/fbxAH2WyAp and set yourself up with a personal assistant
G
GREG ISENBERG @gregisenberg ·
40 reasons 2026 is the best time ever to build a startup
E
eric provencher @pvncher ·
I heard from someone who works at a big tech co that they started rolling out Claude code to employees, with a budget of $100 in credits per month, but people burn through it in 2-3 days. Idk how we scale out agentic work with api pricing
D
David Shapiro (L/0) @DaveShapi ·
85% Of People Will be Unemployable
D
David Shapiro (L/0) @DaveShapi ·
I have revised my estimates for "future employable humans" For reference, my last work estimated around 20% to 25% total labor force participation rate. However, as I've refined my approaches and assumptions, that has been revised down to a LFPR of only 15%. That means that, in the future, I anticipate that less than 1 out of 6 working age adults will have meaningful employment. That may sound abysmal, but the solution is elegant.
D DaveShapi @DaveShapi

85% Of People Will be Unemployable

E
Ejaaz @cryptopunk7213 ·
there it is- "today we're introducing Personal Intelligence" now your emails, photos, youtube & search history, location, documents will all be used to train a personalized version of gemini to deliver you a tailored experience. this is all part of googles multi-pronged masterplan and they're executing much quicker than i expected tbh people are about to realize how powerful their data moat is. openai, anthropic cannot compete. wrote about this in detail here https://t.co/jkShii1XhK
G Google @Google

Today, we’re introducing Personal Intelligence. With your permission, Gemini can now securely connect information from Google apps like @Gmail, @GooglePhotos, Search and @YouTube history with a single tap to make Gemini uniquely helpful & personalized to *you* ✨ This feature is launching in beta today in the @GeminiApp. See Personal Intelligence in action 🧵 ↓

D
Dan ⚡️ @d4m1n ·
since many asked, to "install" all these 1. copy this entire directory: https://t.co/r6fcreGXPZ (including https://t.co/wtrWrWPVid) 2. paste inside the .claude/skills directory in your project 👉 skills only take a bit of context and are loaded when needed by the agent
O
Or Hiltch @_orcaman ·
Today we are launching @openwork_ai, an open-source (MIT-licensed) computer-use agent that’s fast, cheap, and more secure. @openwork_ai  is the result of a short two-day hackathon our team decided to hack, which brings together some of our favorite open source AI modules into one powerful agent, to allow you to: 1. Bring your own model/API key (any provider and model supported by @opencode is supported by Openwork) 2. ~4x faster than Claude for Chrome/Cowork, and much more token-efficient, powered by dev-browser by @sawyerhood (legend) 3. More secure - contrary to Claude for Chrom/Cowork, does not leverage the main browser instance where you are logged into all services already. You login only to the services you need. This significantly reduces the risk of data loss in case of prompt injections, to which computer-use agents are highly exposed. 4. Free and 100% open-source! You can download the DMG (macOS only for now) or fork the github repo via the link in bio (@openwork_ai). Let us know what you think (or better, send a pull request)!
C claudeai @claudeai

Introducing Cowork: Claude Code for the rest of your work. Cowork lets you complete non-technical tasks much like how developers use Claude Code. https://t.co/EqckycvFH3

T
Thariq @trq212 ·
Tool Search now in Claude Code
E
Ethan Mollick @emollick ·
Teaching an experimental class for MBAs on “vibefounding,” the students have four days to come up and launch a company. More on this eventually, but quick observations: 1) I have taught entrepreneurship for over a decade. Everything they are doing in four days would have taken a semester in previous years, if it could have done it at all. Quality is also far better. 2) Give people tools and training and they can do amazing things. We are using a combination of Claude Code, Gemini, and ChatGPT. The non-coders are all building working products. But also everyone is doing weeks of high quality work on financials, research, pricing, positioning, marketing in hours. All the tools are weird to use, even with some training, but they are figuring it out. 3) People with experience in an industry or skill have a huge advantage as they can build solutions that have built-in markets & which solve known hard problems that seemed impossible. (Always been true, but the barriers have fallen to actually doing stuff) 4) The hardest thing to get across is that AI doesn’t just do work for you, it also does new kinds of work. The most successful efforts often take advantage of the fact that the AI itself is very smart. How do you bring its analytical, creative, and empathetic abilities to bear on a problem? What do you do with access to a very smart intelligence on demand? I wish I had more frameworks to clearly teach. So many assumptions about how to launch a business have clearly changed. You don’t need to go through the same discovery process if you build a dozen ideas at the same time & get AI feedback. Many, many new possibilities, and the students really see how big a deal this is.
D
Damian Player @damianplayer ·
this role will become a key hire for most orgs. if you aren’t actively looking for an AI partner, automation specialist, or bringing AI teams in house, you’re already behind. we’re talking to companies doing $5M-$50M/year right now. the demand is insane.
C Codie_Sanchez @Codie_Sanchez

Best money I've ever spent as a CEO... an internal AI transformation hire. He doesn't care about title. He just wants to ship. And he goes across your entire org, sales, revenue, hr, apps, tech and kills stupid manual processes. Such an underrated unlock.

D
Dan Guido @dguido ·
.@trailofbits released our first batch of Claude Skills. Official announcement coming later. https://t.co/vI4amorZrc
M
Miles Deutscher @milesdeutscher ·
If you're building with Claude Code, you'll want to bookmark this site. A full agent marketplace of 60,000+ Claude Skills that are ready for use now. https:// skillsmp. com/ https://t.co/YfZRf4w9TJ
A
Arvind Jain @jainarvind ·
Love this. At @glean, we call these AI Outcomes Managers. They not only lead our internal “Glean on Glean” initiatives, they also work directly with customers to identify high-friction workflows, automate repetitive steps, and deploy AI agents that drive clear business impact.
C Codie_Sanchez @Codie_Sanchez

Best money I've ever spent as a CEO... an internal AI transformation hire. He doesn't care about title. He just wants to ship. And he goes across your entire org, sales, revenue, hr, apps, tech and kills stupid manual processes. Such an underrated unlock.

ℏεsam @Hesamation ·
the Cursor team released a blog post on the best practices of coding with agents. writing fully functional code vs slop comes down to following 10 very simple principles: 1. use plan mode before any code 2. start fresh conversations when it gets confused 3. let the agent get its context, don’t tag everything 4. revert and refine instructions rather than fixing hopelessly 5. add rules for repeated mistakes 6. write tests first so it can iterate 7. run multiple models and pick the best 8. use debug mode for stubborn bugs 9. specific prompts get way better results 10. give it linters and tests to verify​​​​​​​​​​​​​​​​ blog: https://t.co/M9dWf27F4V
B
Boris Cherny @bcherny ·
Super excited about this launch -- every Claude Code user just got way more context, better instruction following, and the ability to plug in even more tools
T trq212 @trq212

Tool Search now in Claude Code

A
Arlan @arlanr ·
it happened mcp is no longer bs
T trq212 @trq212

Tool Search now in Claude Code

A
Alex Sidorenko @asidorenko_ ·
"How can I use react-best-practices skills?" Codex example 👇 https://t.co/dUrnqOUWIu
R rauchg @rauchg

We're encapsulating all our knowledge of @reactjs & @nextjs frontend optimization into a set of reusable skills for agents. This is a 10+ years of experience from the likes of @shuding, distilled for the benefit of every Ralph https://t.co/2QrIl5xa5W

S
Simon Willison @simonw ·
This is great - context pollution is why I rarely used MCP, now that it's solved there's no reason not to hook up dozens or even hundreds of MCPs to Claude Code
T trq212 @trq212

Tool Search now in Claude Code

🎭
🎭 @deepfates ·
Oh you can just make claude code a RLM by telling it to look at its own conversation logs
V
Vercel @vercel ·
We just released 𝚛𝚎𝚊𝚌𝚝-𝚋𝚎𝚜𝚝-𝚙𝚛𝚊𝚌𝚝𝚒𝚌𝚎𝚜, a repo for coding agents. React performance rules and evals to catch regressions, like accidental waterfalls and growing client bundles. How we collected them and how to install the skill ↓ https://t.co/kfLSbKl15X
P
Peter Steinberger @steipete ·
I still think https://t.co/fz1tUJADRo is a better approach. agents know really well how to handle clis.
T trq212 @trq212

Tool Search now in Claude Code

A
am.will @LLMJunky ·
@clawdbot is utterly cracked. From my phone, I had to do repo research, indexing all the migrations, edge functions, and tables I have in my supabase It then passed this context into a Codex agent which used Context7 to pull documentation to help migrate Supabase over to @convex Codex completed the plan, saved it to my repo, and it's ready for migration. Keep in mind, this is a task I could have kicked off from the Denny's parking lot. And to implement the plan? Would have been as instructing it to spin up another Codex (or Claude, Gemini, whatever). @steipete is a legend. I'm only just scratching the surface, but there's an entire library of skills that I've installed. Browser automation, remind me, deep research. It understands images. I can leave it voice memos. I don't know if this is AGI, but its about as close as you can get right now.
G
Guillermo Rauch @rauchg ·
Glimpse of a world of fully generative interfaces. AI → JSON → UI: https://t.co/BKcvtDky5K https://t.co/QH6ctR1ldA
A
Ashpreet Bedi @ashpreetbedi ·
AI Engineering has a Runtime Problem
J
Jeffrey Emanuel @doodlestein ·
If you don’t want to dive directly into my entire Flywheel system all at once, at least try this: 1. Install agent mail using the curl | bash one-liner: curl -fsSL "https://t.co/4cpumwIS41 +%s)" | bash -s -- --yes That will automatically install beads if you don’t already have it. Then install beads_viewer with its one-liner: curl -fsSL "https://t.co/OETEyjZZhN +%s)" | bash Then set up your AGENTS dot md file for your project. You can start with this one and just remove the sections for the tools you’re not using yet: https://t.co/UEViYk7x3Z Then ask CC to adapt it to better fit the tech stack for your particular project. That’s all you need to get started. Then follow this workflow: https://t.co/xkxAQzMPQl Try to start with a smaller, self-contained greenfield (new) project and see whether you can get it all working perfectly without looking at any of the code, just from following the workflow. Spend most of your energy and human time/focus on the markdown plan. Don’t be lazy about the plan! The more you iterate on it with GPT Pro and layer in feedback from other models, the better your project will turn out. Also don’t be lazy about turning the markdown plan into beads, either. Don’t try to one-shot it with CC, you will 100% miss stuff from the plan. This is the easiest thing to screw up assuming you already have a great markdown plan. Do at least 3 rounds of polishing, improving, and expanding the beads. Once you have the beads in good shape based on a great markdown plan, I almost view the project as a foregone conclusion at that point. The rest is basically mindless “machine tending” of your swarm of 5-15 agents as they build out the beads. It’s mostly just juggling these tasks: - Making sure to make them read AGENTS dot md after compactions. - Using many rounds of the “fresh eyes” review prompt whenever an agent tells you it’s done implementing one of the beads. - Swapping accounts when you run out of usage (ugh!). - Making sure you commit frequently to GitHub using my “logically grouped” commits prompt. - When all beads are complete, doing many rounds of the random code inspection and review. - Adding more and more unit and e2e tests. - Setting up gh actions for testing, builds, tags, releases, checksums, etc. - Writing a README and help/docs/tutorials. - Iterating on a “robot mode” (you added one, right?) with feedback from the agents to make it better. - Seeing if you can make your project work better when controlled by Claude Code by making a skill for it. But most of these things can be done using very little mental focus or attention/energy. Save all of that for the ideation and planning phases! The one thing people seem to get wrong is ignoring what I say about planning or transforming their plan into beads. They make a slipshod plan all at once with Claude Code. Or they try to one-shot turning the plan into beads. Or they even do both of those things! Well, of course the project is going to suck and be a buggy mess if you do that. So don’t be lazy. Or if you insist on being lazy, save it for the stages after planning. A great set of beads is all you need. As for the rest of my tools: Once you get comfortable with that workflow, start layering in the other tools, starting with ubs to help find bugs during the review phases. Then add in dcg. You’ll actually appreciate dcg a lot more once Claude wipes out all the work from the other agents since the last commit! As you build up a good session history, layer in cass so you can tap into that history. And then try cm (cass memory system) to start extracting and codifying lessons from your past sessions. And I know I’ve said that I don’t really use ntm yet (I’m not dogfooding it at least), but that’s not quite true. I’ve been using it as a handy building block because of its robot mode. For example, ntm is used by ru (repo_updater) to automate handling gh issues. Good luck, and come to the Discord with any questions!
C craigvandotcom @craigvandotcom

So what would you recommend to someone who wants to start using your stack? I don’t want to use it all at once because then I don’t really feel how it works, if I add layers as I’m comfortable then I’ll feel better. What would be the simple to complex or critical to optional setup sequence?

M
Mario Zechner @badlogicgames ·
Opus is kinda like that excited puppy dog, that will do anything for a belly rub immediately. A messy, cute moron. Codex is like an old donkey that needs some ass kicking to do anything. But once it's going, it's going. In fact, it's hard to stop. Also a messy moron.
M
Muratcan Koylan @koylanai ·
These 17 security Skills for Claude Code are really well-written. - Decision trees agents can actually follow - Authoritative sources with specific file paths - Nested references for deeper context My take is that this is the beginning of something massive. Trail of Bits works with DARPA and Facebook. They don't do things casually. Every company with technical docs will ship Skill packages, not because it's nice to have, but because agents won't adopt your product without them. Agents (or humans) won't read docs; they execute Skills. If you're thinking about how agent-readable knowledge should be structured or are building/leading a startup that plans to create your own Skills: I'd love to chat for 5-10 min to exchange ideas. DMs open.
D dguido @dguido

.@trailofbits released our first batch of Claude Skills. Official announcement coming later. https://t.co/vI4amorZrc

J
Jamon @jamonholmgren ·
40%+ faster runtime … React Native gets even better! Never get tired of seeing this sort of thing even after ten years.
T thymikee @thymikee

Hermes V1 will ship as the default in React Native 0.84 for both iOS and Android. This means: • 2-8% faster startup time • 40%+ faster runtime • faster Metro compilation (less Babel transforms) Just landed in 0.84.0-rc.1 https://t.co/fnH0aMgQxD