M. Semih Babacan
Agentic Coding

Superpowers: The Skill System That Makes Claude Unstoppable

Mehmet Semih Babacan
Mehmet Semih Babacan
March 4, 2026·5 min read

1 views

Observation

Jesse Vincent built something that shouldn't work. A system for structuring prompts into composable modules, with hooks that inject behaviors automatically, slash commands that invoke carefully crafted prompts, and subagents that audit other subagents. It sounds overengineered. It sounds like the kind of thing that falls apart in practice.

But the community didn't just adopt it. They got it. And the reason tells you something important about where prompting is going.

Underlying Mechanism

Here's what Superpowers actually is.

A skill is a markdown file with YAML frontmatter. The frontmatter declares metadata — what the skill is called, when it should activate, whether it's a command or a hook or a background behavior. The body is the prompt itself. That's the whole format. Ridiculously simple. Almost offensively so.

The composition model is where it gets interesting.

Session hooks inject themselves into every conversation automatically. You don't invoke them. They're always there, like gravity. Instructions like "always think step by step before editing code" or "check for regressions before declaring done." The kind of thing you'd normally stuff into a soul.md, except now it's modular. Add one, remove one, swap one out. They don't all compete for the model's attention in one giant file.

Slash commands activate on demand. Type /brainstorm and the corresponding skill fires. This is where Superpowers becomes practical. Instead of reconstructing the perfect prompt from memory every time, you just type a word. The skill contains the prompt you spent twenty minutes crafting. Now it's reusable.

Subagents are the wild part. A skill can spawn a second Claude instance to review the first one's work. Write code, then have a separate agent audit it with different instructions. Two-stage review, automated. The first agent is creative and fast. The second is skeptical and thorough. Sounds like overkill until you try it and realize it catches bugs that would slip through.

There's also a dispatch system with overrides. Personal skills shadow shared skills. Project skills shadow personal skills. Same specificity layering as soul.md, but applied to behaviors instead of rules.

The Technical Patterns That Make It Work

Most people write system prompts like suggestions. "Please consider using best practices." This is like posting a speed limit sign in a town with no cops. The model reads it, nods politely, and does whatever its training distribution suggests.

Superpowers works because it treats prompt engineering as mechanism design. You're not asking the model to be better. You're building structures that make it hard to be worse.

Aggressive Skill Matching: The rule is "If there's even a 1% chance a skill applies, you MUST invoke it." The failure mode isn't false positives—it's the model rationalizing why a skill isn't needed for a task that seems "too simple." There's a Red Flags table with patterns to catch: "This is too simple, no skill needed." You're inoculating against rationalization before it happens.

Two-Stage Subagent Review: When generating code, it spawns a subagent with different instructions to audit the output. The writing prompt says "implement this." The review prompt says "find everything wrong." Different instructions produce genuinely different attention patterns.

TDD as Structural Enforcement: Instructions to delete code and start over if tests aren't written first. This forces the model to specify behavior before generating it.

Skill Chaining and Shadowing: Skills trigger other skills. A coding skill invokes testing, which invokes review. Layering means personal overrides shared, project overrides personal. CSS specificity for agent behavior.

Grounding Enforcement: Every claim cites file:line. Every inference gets labeled. You're forcing the model to show its work.

The common thread: none of these patterns trust the model's judgment about when to apply them. The model's judgment about what to do is often good. Its judgment about when to try hard is terrible. So you take that decision away from it. You build the cops, not just the speed limit signs.

Implication

Why did the community adopt this so aggressively?

Because it makes prompting systematic. Most people prompt AI the way they write emails — off the cuff, inconsistent, wildly varying quality depending on how much coffee they've had. Superpowers forces you to think about your prompting once, get it right, save it, reuse it. The same discipline that made programmers write functions instead of copying code.

But there's a deeper reason. Superpowers is what happens when you take the soul.md idea seriously. Not as a single file, but as a system.

Once you accept that configuring AI happens through language, you want what programmers always want: modularity, composition, overrides, reuse. Superpowers delivers that. It's soul.md evolved into its logical form.

The implication for tool builders is clear. The next generation of agent platforms won't compete on model access. They'll compete on how well they help users articulate identity and behavior. The hard part isn't connecting an agent to your tools. The hard part is defining what kind of agent should use those tools.

Clean Stop

The interesting question isn't whether Superpowers wins. It's what comes after. More skill systems. Better composition. Maybe AI-authored skills that learn from your corrections.

What started as a markdown file in a repo has become a discipline. And disciplines have momentum.

This is part of a series on identity engineering for AI agents. See also: "Your Soul Is a Markdown File Now" — the post that started it all.