Karpathy Builds an LLM Council While Claude Code Power Users Share Their Best Hooks and Skills
Daily Wrap-Up
The dominant thread today was developer tooling maturity. Five separate posts shared configurations, skills, hooks guides, and prompts for Claude Code and GitHub Copilot, which signals that the coding assistant ecosystem has moved well past the "is this useful?" phase into "how do I squeeze every last drop of value out of this?" territory. The most interesting tension sits between @github's admission that Copilot code review is non-deterministic and the community's response, which is essentially to paper over that randomness with increasingly detailed instruction files and configuration. It's prompt engineering all the way down, but the practitioners sharing their setups today are clearly getting real results from the investment.
On the agent side, @clare_liguori revealed that Amazon has thousands of internal agent SOPs (Standard Operating Procedures) for coding assistants and just open-sourced the format. That's a significant signal about where enterprise automation is heading. Meanwhile, @NoahEpstein_ showed off an MCP that connects directly to n8n instances, building and debugging workflows autonomously. These two posts together paint a picture of agents moving from novelty demos to structured, repeatable operational tooling. The creative generation posts were the fun counterpoint: @karpathy spent his Saturday vibe-coding an "LLM council" app, and multiple people were producing beautiful exploded engineering diagrams of retro hardware with Nano Banana Pro. Sometimes the best showcase for AI capability is just making something cool on a weekend.
The most entertaining moment was @vasuman's honest critique of vibe-coding culture, calling out the "fake and lame energy" of people claiming you can build a million-dollar app in a few days. It's a healthy corrective to the hype cycle, and it landed harder because it came from someone who admits they're part of the problem. The most practical takeaway for developers: invest time in your coding assistant configuration files. Whether it's Copilot's instructions file, Claude Code hooks and skills, or Codex's config.toml, the posts today consistently showed that structured guidance dramatically improves output quality, and the gap between a default setup and a tuned one is widening fast.
Quick Hits
- @tom_doerr shared a self-hosted long-term memory system for AI, which is increasingly relevant as agents need persistent context across sessions.
- @saltyAom recommended Better Auth as a drop-in authentication library, praising it for removing redundant setup and letting you get straight to building.
- @_vmlops highlighted a Stanford AI paper that's making the rounds, though the post was light on specifics about what makes it worth reading.
- @danielhangan_ claims TikTok's algorithm weights an "IP address trust score" based on analysis of 847 accounts, which, if true, would explain a lot about inconsistent reach on the platform.
- @dkundel dropped a reply link without much context, seemingly sharing a resource in response to a question.
- @lindynap posted a sarcastic take on hiring and human capital, referencing a comparison that didn't land with much nuance.
Coding Assistants Level Up: Hooks, Skills, and Configuration as Competitive Advantage
The coding assistant ecosystem is quietly undergoing a configuration renaissance. Today's feed featured no fewer than five posts from practitioners sharing their custom setups, and the collective message is clear: the default experience is just the starting line.
@dani_avila7 published a comprehensive guide to Claude Code hooks, noting that "the hook system is incredibly powerful, but the docs don't really explain when to use each one." The guide breaks down which hook to use for which scenario, addressing a real gap in the official documentation. In a similar vein, @svpino shared a custom Claude Code skill for generating commit messages, explaining that "this improves commit messages significantly. It also prevents Claude from including a 'Generated with Claude Code' disclaimer on every commit message." These are the kinds of quality-of-life customizations that compound over time into meaningfully different workflows.
On the Copilot side, @github acknowledged a fundamental challenge with AI code review: non-determinism. Their recommendation is to treat the instructions file as the primary lever for consistency, sharing practical tips and common pitfalls drawn from real-world usage data. @nummanali went full YOLO with an OpenAI Codex configuration, highlighting the new streamable_shell = true setting that enables interactive shell mode where "the agent can run shells in the background" and "send keystrokes to interact with options." And @donvito offered a simpler but effective contribution: a single prompt for Claude Code that reportedly "10x levels up" website design output.
What ties all of these together is the realization that AI coding tools are becoming platforms with their own configuration languages, extension systems, and community-shared best practices. The developers who invest in understanding these systems at the configuration level are pulling ahead of those who just type prompts into a chat window. Meanwhile, @vasuman offered a timely counterpoint to the hype, observing that vibe-coding "opened the floodgates to a certain kind of person who now pushes the idea that you can vibe-code an app in a few days and start printing life changing amounts of money." The critique resonated precisely because it came from a self-aware participant rather than a skeptic on the outside.
Agents Graduate From Demos to Operating Procedures
The agent conversation has shifted. Instead of "look what this agent can do," today's posts focused on structure, patterns, and repeatability, which is exactly what you'd expect as the technology moves toward production use.
The biggest signal came from @clare_liguori at Amazon, who shared a new open-source project called Agent SOPs: "We use this structured agent prompt format a LOT at Amazon with coding assistants to automate daily work. There are 1000s of agent SOPs internally." The fact that Amazon has thousands of these internally and is now open-sourcing the format suggests that agent orchestration is becoming a first-class engineering discipline rather than an ad-hoc experiment. On the architectural side, @Saboo_Shubham_ broke down the "Parallel Fan-out Gather Agent Pattern," one of the core agentic design patterns where work is distributed across multiple agents and results are collected. It's the kind of pattern that separates toy demos from systems that actually scale.
The applied examples were equally telling. @NoahEpstein_ demonstrated an MCP that connects to n8n workflow instances, describing a loop where the agent "builds the workflow, deploys it to YOUR n8n, runs it, watches it fail, debugs it, fixes it, runs it again." That autonomous debug-fix-retry loop is where agents start to deliver genuine leverage. @hamza_automates built a restaurant agent in ten minutes that handles "customer inquiries, reservation flow, menu questions, upsells, follow-up reviews," claiming it "solved a problem they've been struggling with for years." The speed of deployment matters less than the fact that the problem space (structured customer interactions with predictable flows) is exactly where agents work well today. The pattern across all four posts is convergence on structured, repeatable agent workflows rather than open-ended autonomy.
Models Get Creative: LLM Councils, Animation Engines, and Retro Hardware Art
Today's model-focused posts split between practical multi-model orchestration and pure creative exploration, and both directions revealed interesting things about where capabilities are heading.
@karpathy's weekend project was characteristically elegant: an "llm-council" web app that "looks exactly like ChatGPT except each user query is dispatched to multiple models on your council using OpenRouter." The idea of routing queries to multiple models and synthesizing responses isn't new, but Karpathy packaging it as a clean web app makes the concept accessible. @stevenbjohnson picked up the thread and ran with it, creating "a NotebookLM notebook based on this tweet, and then did a Deep Research run in-app to gather related sources. Then generated one of our new slide decks to explore further. Instant knowledge base." It's a nice demonstration of how different AI tools can chain together: one person's weekend hack becomes another person's research starting point within hours.
On the visual generation side, @MengTo made a bold claim: "Gemini 3 is the best model at creating animations. It's not even close." The post focused on UI animations specifically, noting key placement recommendations for "the hero intro, hover interactions, slow transitions." Meanwhile, Nano Banana Pro had a strong showing with two separate posts praising its output. @sahilypatel generated "exploded engineering diagrams of retro devices" including a Nintendo Gameboy, Sony Walkman, iPod Classic, and Polaroid SX-70, marveling at how "crazy how beautiful hardware used to be." @egeberkina echoed the enthusiasm, noting the model "structured the whole layout perfectly" from a single prompt. The visual generation space is fragmenting in an interesting way: rather than one model dominating everything, specific models are carving out niches in animation, technical illustration, and layout, which gives developers and designers real choices based on their specific use case.