prompt-library

Insights

Higher-level findings computed from your session logs. Every bullet is an SQL query over your data — no AI interpretation, just numbers you can verify yourself. Read these as hypotheses about how you prompt, not rules.

Your length signal

Your golden prompts have a median of 89 chars vs 71 for all prompts — 1.3× longer.

The prompts that unlock long autonomous runs are meaningfully longer than your baseline. Terseness isn't what's getting you leverage — specificity is.

median chars, 536 golden vs 7,392 total prompts

"critique" works for you

When you include "critique" in a prompt, correction rate drops 100% — 0.0% vs your 11.4% baseline.

This is a meta-instruction you should keep using. Other meta-instructions with strong effects are shown below.

n=11 prompts containing "critique"

Session drift

Your correction rate climbs from 6.6% in the first 3 turns to 11.9% after turn 26 — drift is real.

You start sessions crisp but grow more corrective deeper in. Two plausible fixes: shorter sessions, or a mid-session reset ritual (summarize, re-plan, clear stale context).

Turn 1-3: 380 prompts • Turn 26+: 5,258 prompts

Sweet spot for prompt length

Prompts in the 800+ char range are 5.8× more likely to become golden than your shortest ones.

Golden rate by length: <40: 2.3% · 40-120: 9.5% · 120-300: 9.5% · 300-800: 7.1% · 800+: 13.5%

384 prompts in the winning bucket

Where you flow vs. where you fight

code/remote-browser-sandbox has your cleanest convergence (94%); sandbox has your worst (83%).

Big gaps across projects usually mean something environmental: unfamiliar codebase, bad tooling, missing context docs, or fatigue. Worth asking what the clean projects have that the messy ones don't.

Best: 3 sessions · Worst: 11 sessions (min 8 user turns each)

Recent trajectory

Your convergence is holding steady — last 2 weeks averaged 89.5%, earlier weeks averaged 88.6%.

Convergence is the share of your user turns that didn't trigger a correction. Trend over time tells you whether your prompting is improving.

6 weeks of data

Session openers that just worked

5 recent sessions started with no corrections at all, for 6+ user turns.

These are the opening prompts. Steal from yourself.

Perfect-convergence sessions with ≥6 user turns

Openers from your best sessions

  • need you to monitor for changes and for the next 4 hours, push them to the remote when they're in a good state, monitor vercel deployment (u have vercel cli its opendatalabs/macro-cart project)... other agents are actively working but forget to push their work
    59 user turns, no corrections
  • hey right now my devcontainers load ~/code as /projects. is there anyway to preserve the path the way we do for other stuff? investigate until you fully understand my question
    25 user turns, no corrections
  • Connect my United (airline) data. Skill: https://github.com/vana-com/data-connectors/tree/main/skills/vana-connect. work exclusively within /tmp/united and do not even look at any other code in this environment.
    14 user turns, no corrections
  • ~/clawd/research/voxtral-tts-setup.md fetch that from [email protected]
    9 user turns, no corrections
  • Clone https://github.com/vana-com/data-connectors.git (branch feat/agent-connect-skill), read skills/vana-connect/SKILL.md, and follow its instructions to connect my Linear data. Do not use API keys — the connector must use browser login.
    7 user turns, no corrections

Meta-instruction effects

Phrasencorr. withcorr. withouteffect
critique110.0%11.4% 11.4 pts
ask me80.0%11.4% 11.4 pts
verify515.9%11.4% 5.5 pts
research2367.6%11.5% 3.9 pts
prior art378.1%11.4% 3.3 pts
ultrathink119.1%11.4% 2.3 pts
root cause2015.0%11.4% 3.6 pts
why37616.5%11.1% 5.4 pts
think hard1136.4%11.4% 25.0 pts