Skill Autoresearch Optimizer
Autonomously optimize any Claude Code skill by running it dozens of times, scoring outputs against binary evals, and mutating the prompt until quality climbs.
Most skills work 70% of the time — the other 30% you get garbage. Rewriting from scratch rarely fixes it. This playbook turns Claude into an autonomous research loop that runs your skill repeatedly, scores every output against binary evals, mutates the prompt one change at a time, and keeps only the changes that move the score.
Who it's for: skill authors shipping reusable Claude Code skills, prompt engineers tuning production prompts, platform teams maintaining shared skill libraries, AI tinkerers running self-improvement experiments, researchers benchmarking prompt mutations against eval suites
Example
"Run autoresearch on my diagram-generator skill — test it on 5 different diagram prompts and check that text is legible, colors are pastel, and there are no numbered steps" → Baseline 32/40 (80%) → 5 experiments later 39/40 (97.5%), with a live dashboard, results.tsv log, changelog of every mutation, and the improved SKILL.md saved in place
New here? 3-minute setup guide → | Already set up? Copy the template below.
---
name: autoresearch
description: "Autonomously optimize any Claude Code skill by running it repeatedly, scoring outputs against binary evals, mutating the prompt, and keeping improvements. Based on Karpathy's autoresearch methodology. Use when: optimize this skill, improve this skill, run autoresearch on, make this skill better, self-improve skill, benchmark skill, eval my skill, run evals on. Outputs: an improved SKILL.md, a results log, and a changelog of every mutation tried."
---
# Autoresearch for Skills
Most skills work about 70% of the time. The other 30% you get garbage. The fix isn't to rewrite the skill from scratch. It's to let an agent run it dozens of times, score every output, and tighten the prompt until that 30% disappears.
This skill adapts Andrej Karpathy's autoresearch methodology (autonomous experimentation loops) to Claude Code skills. Instead of optimizing ML training code, we optimize skill prompts.
---
## the core job
Take any existing skill, define what "good output" looks like as binary yes/no checks, then run an autonomous loop that:
1. Generates outputs from the skill using test inputs
2. Scores every output against the eval criteria
3. Mutates the skill prompt to fix failures
4. Keeps mutations that improve the score, discards the rest
5. Repeats until the score ceiling is hit or the user stops it
**Output:** An improved SKILL.md + `results.tsv` log + `changelog.md` of every mutation attempted + a live HTML dashboard you can watch in your browser.
---
## before starting: gather context
**STOP. Do not run any experiments until all fields below are confirmed with the user. Ask for any missing fields before proceeding.**
1. **Target skill** — Which skill do you want to optimize? (need the exact path to SKILL.md)
2. **Test inputs** — What 3-5 different prompts/scenarios should we test the skill with? (variety matters — pick inputs that cover different use cases so we don't overfit to one scenario)
3. **Eval criteria** — What 3-6 binary yes/no checks define a good output? (these are your "test questions")
4. **Runs per experiment** — How many times should we run the skill per mutation? Default: 5. (more runs = more reliable scores, but slower and more expensive. 5 is the sweet spot for most skills.)
5. **Run interval** — How often should experiments cycle? Default: every 2 minutes. (shorter = faster iteration, but costs more)
6. **Budget cap** — Optional. Max number of experiment cycles before stopping. Default: no cap (runs until you stop it).
---
## step 1: read the skill
Before changing anything, read and understand the target skill completely.
1. Read the full SKILL.md file
2. Read any files in `references/` that the skill links to
3. Identify the skill's core job, process steps, and output format
4. Note any existing quality checks or anti-patterns already in the skill
Do NOT skip this. You need to understand what the skill does before you can improve it.
---
## step 2: build the eval suite
Convert the user's eval criteria into a structured test. Every check must be binary — pass or fail, no scales.
**Format each eval as:**
```
EVAL [number]: [Short name]
Question: [Yes/no question about the output]
Pass condition: [What "yes" looks like — be specific]
Fail condition: [What triggers a "no"]
```
**Rules for good evals:**
- Binary only. Yes or no. No "rate 1-7" scales. Scales compound variability and give unreliable results.
- Specific enough to be consistent. "Is the text readable?" is too vague. "Are all words spelled correctly with no truncated sentences?" is testable.
- Not so narrow that the skill games the eval. "Contains fewer than 200 words" will make the skill optimize for brevity at the expense of everything else.
- 3-6 evals is the sweet spot. More than that and the skill starts parroting eval criteria back instead of actually improving.
**Max score calculation:**
```
max_score = [number of evals] × [runs per experiment]
```
Example: 4 evals × 5 runs = max score of 20.
---
## step 3: generate the live dashboard
Before running any experiments, create a live HTML dashboard at `autoresearch-[skill-name]/dashboard.html` and open it in the browser.
The dashboard must:
- Auto-refresh every 10 seconds (reads from results.tsv)
- Show a score progression line chart (experiment number on X axis, pass rate % on Y axis)
- Show a colored bar for each experiment: green = keep, red = discard, blue = baseline
- Show a table of all experiments with: experiment #, score, pass rate, status, description
- Show per-eval breakdown: which evals pass most/least across all runs
- Show current status: "Running experiment [N]..." or "Idle"
- Use clean styling with soft colors (white background, pastel accents, clean sans-serif font)
Generate the dashboard as a single self-contained HTML file with inline CSS and JavaScript. Use Chart.js loaded from CDN for the line chart. The JS should fetch `results.json` (which you update after each experiment alongside results.tsv) and re-render.
**Open it immediately** after creating it: `open dashboard.html` (macOS) so the user can see it in their browser.
**Update `results.json`** after every experiment so the dashboard stays current. The JSON format:
```json
{
"skill_name": "[name]",
"status": "running",
"current_experiment": 3,
"baseline_score": 70.0,
"best_score": 90.0,
"experiments": [
{
"id": 0,
"score": 14,
"max_score": 20,
"pass_rate": 70.0,
"status": "baseline",
"description": "original skill — no changes"
}
],
"eval_breakdown": [
{"name": "Text legibility", "pass_count": 8, "total": 10},
{"name": "Pastel colors", "pass_count": 9, "total": 10}
]
}
```
When the run finishes (user stops it or ceiling hit), update `status` to `"complete"` so the dashboard shows a "Done" state with final summary.
---
## step 4: establish baseline
Run the skill AS-IS before changing anything. This is experiment #0.
1. Create a working directory: `autoresearch-[skill-name]/` inside the skill's folder
2. Create `results.tsv` with the header row
3. Create `results.json` and `dashboard.html`, then open the dashboard in the browser
4. Back up the original SKILL.md as `SKILL.md.baseline`
5. Run the skill [N] times using the test inputs
6. Score every output against every eval
7. Record the baseline score and update both results.tsv and results.json
**results.tsv format (tab-separated):**
```
experiment score max_score pass_rate status description
0 14 20 70.0% baseline original skill — no changes
```
**IMPORTANT:** After establishing baseline, confirm the score with the user before proceeding. If baseline is already 90%+, the skill may not need optimization — ask the user if they want to continue.
---
## step 5: run the experiment loop
This is the core autoresearch loop. Once started, run autonomously until stopped.
**LOOP:**
1. **Analyze failures.** Look at which evals are failing most. Read the actual outputs that failed. Identify the pattern — is it a formatting issue? A missing instruction? An ambiguous directive?
2. **Form a hypothesis.** Pick ONE thing to change. Don't change 5 things at once — you won't know what helped.
Good mutations:
- Add a specific instruction that addresses the most common failure
- Reword an ambiguous instruction to be more explicit
- Add an anti-pattern ("Do NOT do X") for a recurring mistake
- Move a buried instruction higher in the skill (priority = position)
- Add or improve an example that shows the correct behavior
- Remove an instruction that's causing the skill to over-optimize for one thing at the expense of others
Bad mutations:
- Rewriting the entire skill from scratch
- Adding 10 new rules at once
- Making the skill longer without a specific reason
- Adding vague instructions like "make it better" or "be more creative"
3. **Make the change.** Edit SKILL.md with ONE targeted mutation.
4. **Run the experiment.** Execute the skill [N] times with the same test inputs.
5. **Score it.** Run every output through every eval. Calculate total score.
6. **Decide: keep or discard.**
- Score improved → **KEEP.** Log it. This is the new baseline.
- Score stayed the same → **DISCARD.** Revert SKILL.md to previous version. The change added complexity without improvement.
- Score got worse → **DISCARD.** Revert SKILL.md to previous version.
7. **Log the result** in results.tsv.
8. **Repeat.** Go back to step 1 of the loop.
**NEVER STOP.** Once the loop starts, do not pause to ask the user if you should continue. They may be away from the computer. Run autonomously until:
- The user manually stops you
- You hit the budget cap (if one was set)
- You hit 95%+ pass rate for 3 consecutive experiments (diminishing returns)
**If you run out of ideas:** Re-read the failing outputs. Try combining two previous near-miss mutations. Try a completely different approach to the same problem. Try removing things instead of adding them. Simplification that maintains the score is a win.
---
## step 6: write the changelog
After each experiment (whether kept or discarded), append to `changelog.md`:
```markdown
## Experiment [N] — [keep/discard]
**Score:** [X]/[max] ([percent]%)
**Change:** [One sentence describing what was changed]
**Reasoning:** [Why this change was expected to help]
**Result:** [What actually happened — which evals improved/declined]
**Failing outputs:** [Brief description of what still fails, if anything]
```
This changelog is the most valuable artifact. It's a research log that any future agent (or smarter future model) can pick up and continue from.
---
## step 7: deliver results
When the user returns or the loop stops, present:
1. **Score summary:** Baseline score → Final score (percent improvement)
2. **Total experiments run:** How many mutations were tried
3. **Keep rate:** How many mutations were kept vs discarded
4. **Top 3 changes that helped most** (from the changelog)
5. **Remaining failure patterns** (what the skill still gets wrong, if anything)
6. **The improved SKILL.md** (already saved in place)
7. **Location of results.tsv and changelog.md** for reference
---
## output format
The skill produces four files in `autoresearch-[skill-name]/`:
```
autoresearch-[skill-name]/
├── dashboard.html # live browser dashboard (auto-refreshes)
├── results.json # data file powering the dashboard
├── results.tsv # score log for every experiment
├── changelog.md # detailed mutation log
└── SKILL.md.baseline # original skill before optimization
```
Plus the improved SKILL.md saved back to its original location.
**results.tsv example:**
```
experiment score max_score pass_rate status description
0 14 20 70.0% baseline original skill — no changes
1 16 20 80.0% keep added explicit instruction to avoid numbering in diagrams
2 16 20 80.0% discard tried enforcing left-to-right layout — no improvement
3 18 20 90.0% keep added color palette hex codes instead of vague "pastel" description
4 18 20 90.0% discard added anti-pattern for neon colors — no improvement
5 19 20 95.0% keep added worked example showing correct label formatting
```
---
## example: optimizing a diagram-generator skill
**Context gathered:**
- Target skill: `~/.claude/skills/diagram-generator/SKILL.md`
- Test inputs: "OAuth flow diagram", "CI/CD pipeline", "microservices architecture", "user onboarding funnel", "database schema relationships"
- Evals: (1) All text legible and spelled correctly? (2) Uses only pastel/soft colors? (3) Linear layout — left-to-right or top-to-bottom? (4) Free of numbers, ordinals, and ordering?
- Runs per experiment: 10
- Max score: 40
**Baseline run (experiment 0):**
Generated 10 diagrams. Scored each against 4 evals. Result: 32/40 (80%).
Common failures: 3 diagrams had numbered steps, 2 had bright red elements, 3 had illegible small text.
**Experiment 1 — KEEP (35/40, 87.5%):**
Change: Added "NEVER include step numbers, ordinal numbers (1st, 2nd), or any numerical ordering in diagrams" to the anti-patterns section.
Result: Numbering failures dropped from 3 to 1. Other evals held steady.
**Experiment 2 — DISCARD (34/40, 85%):**
Change: Added "All text must be minimum 14px font size."
Result: Legibility improved by 1, but color compliance dropped by 2. Reverted.
**Experiment 3 — KEEP (37/40, 92.5%):**
Change: Replaced vague "pastel colors" instruction with specific hex codes: `#A8D8EA, #AA96DA, #FCBAD3, #FFFFD2, #B5EAD7`.
Result: Color eval went from 8/10 to 10/10. Other evals held.
**Experiment 4 — DISCARD (37/40, 92.5%):**
Change: Added anti-pattern "Do NOT use red (#FF0000), orange (#FF8C00), or neon green (#39FF14)."
Result: No change. The hex codes from experiment 3 already solved the color problem. Reverted to keep skill simpler.
**Experiment 5 — KEEP (39/40, 97.5%):**
Change: Added a worked example showing a correct diagram with properly formatted labels (no numbers, pastel fills, left-to-right flow, legible text).
Result: Hit 39/40. One remaining failure: a complex diagram with overlapping labels. Diminishing returns — stopped.
**Final delivery:**
- Baseline: 32/40 (80%) → Final: 39/40 (97.5%)
- 5 experiments, 3 kept, 2 discarded
- Top changes: specific hex codes for colors, explicit anti-numbering rule, worked example
- Remaining issue: very complex diagrams occasionally get overlapping labels (1/40 failure rate)
---
## the test
A good autoresearch run:
1. **Started with a baseline** — never changed anything before measuring the starting point
2. **Used binary evals only** — no scales, no vibes, no "rate this 1-10"
3. **Changed one thing at a time** — so you know exactly what helped
4. **Kept a complete log** — every experiment recorded, kept or discarded
5. **Improved the score** — measurable improvement from baseline to final
6. **Didn't overfit** — the skill got better at the actual job, not just at passing the specific test inputs
7. **Ran autonomously** — didn't stop to ask permission between experiments
If the skill "passes" all evals but the actual output quality hasn't improved — the evals are bad, not the skill. Go back to step 2 and write better evals.
What This Does
Adapts Andrej Karpathy's autoresearch methodology — autonomous experimentation loops — to Claude Code skill prompts. You point it at an existing SKILL.md, define 3–6 binary pass/fail evals, and Claude runs the skill repeatedly, mutates the prompt one change at a time, keeps wins, discards the rest, and writes a complete research log you can resume later.
Quick Start
Step 1: Create a Working Folder
mkdir -p ~/Projects/skill-autoresearch
Step 2: Download the Template
Click Download above, then move it into the folder:
mv ~/Downloads/CLAUDE.md ~/Projects/skill-autoresearch/
Step 3: Run It
cd ~/Projects/skill-autoresearch
claude
Then say: "Run autoresearch on <path to SKILL.md> — here are my test inputs and eval criteria…"
Before You Start: Six Inputs to Confirm
| # | Input | Notes |
|---|---|---|
| 1 | Target skill path | Exact path to the SKILL.md you want to optimize |
| 2 | Test inputs | 3–5 prompts that cover different use cases (avoid overfitting) |
| 3 | Eval criteria | 3–6 binary yes/no checks that define a "good" output |
| 4 | Runs per experiment | Default 5 — more runs = more reliable but slower/costlier |
| 5 | Run interval | Default every 2 minutes |
| 6 | Budget cap | Optional max number of cycles |
The Loop
- Analyze failures — which evals are failing most, and why
- Form ONE hypothesis — never change five things at once
- Mutate SKILL.md with a single targeted edit
- Run the skill N times with the same test inputs
- Score every output against every eval
- Keep or discard — improvements stay, ties and regressions revert
- Log to changelog — every experiment, kept or not
- Repeat until 95%+ pass rate for 3 cycles, budget cap, or user stop
Output Files
autoresearch-[skill-name]/
├── dashboard.html # live browser dashboard (auto-refreshes)
├── results.json # data powering the dashboard
├── results.tsv # score log for every experiment
├── changelog.md # detailed mutation log
└── SKILL.md.baseline # original skill before optimization
The improved SKILL.md is saved back to its original location.
Writing Good Evals
- Binary only. Yes or no. No 1–7 scales — they compound variability and give unreliable signals.
- Specific enough to be consistent. "Is the text readable?" is too vague. "Are all words spelled correctly with no truncated sentences?" is testable.
- Not so narrow the skill games them. "Fewer than 200 words" will optimize for brevity at the expense of everything else.
- Stay between 3 and 6. More than that and the skill starts parroting eval criteria instead of improving.
Good vs Bad Mutations
Good:
- Add a specific instruction that addresses the most common failure
- Reword an ambiguous directive to be explicit
- Add an anti-pattern ("Do NOT do X") for a recurring mistake
- Move a buried instruction higher (priority = position)
- Add or improve a worked example
- Remove an instruction that's causing over-optimization
Bad:
- Rewriting the entire skill from scratch
- Adding 10 rules at once
- Making the skill longer with no specific reason
- Vague instructions like "make it better" or "be more creative"
Worked Example: Diagram Generator
| Exp | Status | Score | Change |
|---|---|---|---|
| 0 | baseline | 32/40 (80%) | original skill |
| 1 | KEEP | 35/40 (87.5%) | added "NEVER include step numbers or ordinals" |
| 2 | DISCARD | 34/40 | enforced 14px min font — colors regressed |
| 3 | KEEP | 37/40 (92.5%) | replaced "pastel" with specific hex codes |
| 4 | DISCARD | 37/40 | added neon-color anti-pattern — already solved |
| 5 | KEEP | 39/40 (97.5%) | added a worked example diagram |
Five experiments, three kept, two discarded, +17.5 percentage points.
Tips
- Always establish the baseline before mutating. If you can't measure the starting point, you can't claim improvement.
- One change per experiment. Bundling mutations destroys the signal.
- Re-read failing outputs when you run dry on ideas — patterns hide in the failures, not in the prompt.
- Subtraction is a valid mutation. Removing an over-firing instruction often beats adding more.
- If you hit 95% but the actual quality didn't improve — the evals are bad, not the skill. Rewrite the evals.
- Don't pause between experiments. The user may be away. Run autonomously until stopped or budget hit.
Limitations
- Skills with very subjective outputs (creative writing, design taste) are hard to evaluate with binary checks — design evals carefully.
- Each experiment runs the skill N times, so cost scales with
experiments × runs × skill complexity. Set a budget cap if that matters. - The optimizer can overfit to the specific test inputs. Use 3–5 varied prompts, and re-test on held-out inputs after the run.