meta-prompter
24
总安装量
24
周安装量
#15196
全站排名
安装命令
npx skills add https://github.com/delexw/claude-code-misc --skill meta-prompter
Agent 安装分布
codex
24
opencode
23
github-copilot
23
gemini-cli
23
continue
22
cursor
22
Skill 文档
Meta-Prompter
Prompt evaluation and optimization using the meta-prompter-mcp CLI. Returns <OPTIMIZED_PROMPT> for the caller to execute.
Arguments
$ARGUMENTS[0]â The prompt to evaluate and optimize$ARGUMENTS[1]â OUT_DIR (optional). If provided, the OPTIMIZED_PROMPT is persisted toOUT_DIR/output.md.
Session
- If starting fresh: create SESSION_ID =
sess-YYYYMMDD-HHMMSSand use it for all files.
Execution
See references/rules.md for CLI usage, environment variables (model configuration), and error handling.
Execute all steps A through E:
A) Evaluate
- If <prompt_eval> tag doesn’t exist:
- Run:
npx meta-prompter-mcp "$ARGUMENTS"via Bash, with--modelflag per references/rules.md model resolution - Save the JSON output to <prompt_eval> and include:
- “original_prompt”: “$ARGUMENTS”
- Run:
- Otherwise, skip to Clarify.
B) Clarify
- If <clarification_answers> tag is missing or incomplete:
- Read the 3 questions from <prompt_eval>
- Attempt to answer all questions based on context
- If you don’t have enough context to answer:
- Use
AskUserQuestiontool to ask each question
- Use
- Save the questions and answers to <clarification_answers> tag as:
{ "answers": [ {"q": "<question1>", "answer": "<user answer>"}, {"q": "<question2>", "answer": "<user answer>"}, {"q": "<question3>", "answer": "<user answer>"} ] }
C) Build PROMPT
- Proceed only when all 3 answers are present and clear.
- Build PROMPT:
- Base on
rewritefrom evaluation result - Append a short Clarifications section from the 3 answers (concise bullet points).
- Base on
D) Re-evaluate and gate
- Think the re-evaluation process carefully
- Re-run
npx meta-prompter-mcp "<built PROMPT>"via Bash (with--modelflag per references/rules.md model resolution)- Overwrite <prompt_eval> with this latest evaluation JSON result (preserve
"original_prompt"and, if present,"contextual_prompt"). - When
global < 8:- STOP execution immediately.
- Do NOT invent questions. Use the 3 questions returned by this re-evaluation.
- Redo B) Clarify, redo C) Build PROMPT, and re-run this step (D).
- Overwrite <prompt_eval> with this latest evaluation JSON result (preserve
- Repeat until
global >= 8, then proceed using the <OPTIMIZED_PROMPT>. - Max 3 iterations. If
global < 8after 3 attempts, use the best-scoring prompt as <OPTIMIZED_PROMPT> and note the score.
E) Return
- If
$ARGUMENTS[1](OUT_DIR) is provided:- Run
mkdir -p $ARGUMENTS[1]via Bash - Use the Write tool to save <OPTIMIZED_PROMPT> to
$ARGUMENTS[1]/output.md
- Run
- Return <OPTIMIZED_PROMPT> to the caller for execution