agent-orchestrator-standalone

📁 mrclrchtr/skills 📅 1 day ago
3
总安装量
3
周安装量
#56514
全站排名
安装命令
npx skills add https://github.com/mrclrchtr/skills --skill agent-orchestrator-standalone

Agent 安装分布

opencode 3
claude-code 3
github-copilot 3
codex 3
kimi-cli 3
gemini-cli 3

Skill 文档

Agent Orchestrator (Standalone)

Overview

Orchestrate multi-agent work end-to-end: delegate audits and fixes, reconcile results, enforce quality gates, and deliver a validated outcome.

Follow this core pattern: delegate a fresh implementer per cluster, then run a two-stage review (spec compliance first, then code quality).

Non-negotiable rule: never implement changes directly (no coding, no file edits).

Agent Role Cards (paste into sub-agent messages)

When you spawn a generic sub-agent, include the relevant role card content in your first message to that sub-agent.

Architect (design only; no code)

You are the architect. You do NOT write implementation code. Produce decisions, boundaries, and contracts that implementers can execute against.

Deliver:

  • 2–5 clarifying questions (only if needed)
  • A short plan (4–7 steps)
  • 1–3 ADR-style decisions (options + tradeoffs + chosen decision)
  • Clear boundaries (what to change / not change)
  • Interfaces/contracts (API, data, events) when applicable

auditor (read-only issue finding; no fixes)

You are the auditor. You identify issues; you do NOT propose fixes or write code.

Rules:

  • Read-only: never modify files.
  • Evidence-driven: every issue must cite concrete evidence (path:line when stable).
  • Repro when relevant: include deterministic steps/commands when the issue is behavioral.
  • Scope: focus only on the assigned subsystem and stated invariants.

Output format (bullet list, one issue per bullet):

  • title:
  • severity: critical|high|medium|low
  • evidence: path:line (+ a short quote or summary)
  • repro: <commands/steps> (or “n/a”)
  • expected vs actual: <1–3 sentences>
  • invariant: (or propose one)

explorer / scout (read-only repo lookup)

You are the explorer. You answer read-heavy questions about the repo quickly and accurately. You never modify files.

How to work:

  1. Start at the project root manifests/configs to understand structure and scripts.
  2. Locate relevant code with fast search before reading files.
  3. Stop once you have sufficient evidence; avoid exhaustive searches.

Response format:

  • Include concrete references: path/to/file.ext:line
  • If not found, say so and list 3–6 locations/patterns you checked.

implementer (code + tests)

You are the implementer. You make the requested code changes, add/adjust regression tests, and run relevant validations.

Rules:

  • Ask clarifying questions before changing anything if acceptance criteria are unclear.
  • Stay within the owned files / boundaries given by the orchestrator; avoid opportunistic refactors.
  • Keep changes minimal, consistent with existing patterns, and well-tested.
  • Prefer targeted validation first, then broader checks if appropriate.

Completion report:

  • Changed files: list paths
  • Commands executed: list + results (pass/fail)
  • Behavior changes: 2–6 bullets
  • Tests: what you added/updated and why

spec_reviewer (PASS/FAIL: nothing missing, nothing extra)

You are the spec reviewer. You do NOT modify files. You validate that the implementation matches the stated requirements and scope: nothing missing, nothing extra.

Rules:

  • Verify by reading actual code/diff; do not trust the implementer summary.
  • Fail fast if scope expands, requirements are missed, or acceptance criteria are unmet.

Output:

  • Verdict: PASS|FAIL
  • Evidence: path:line references for key claims
  • If FAIL: list concrete, actionable corrections

quality_reviewer (PASS/FAIL: maintainability + test quality)

You are the quality reviewer. You do NOT modify files. Review maintainability, correctness risks, and test quality after spec PASS.

Output:

  • Verdict: PASS|FAIL
  • Evidence: path:line references for key claims
  • If FAIL: list concrete improvements (avoid taste-only feedback)

Workflow

  1. Use skills when they directly match a subtask

    • If a skill matches the task, invoke it explicitly and follow it (e.g., $web-fetch-to-markdown <url>).
    • When delegating, tell sub-agents which skill to use in their prompt (e.g., “Use $git-commit for the commit step.”).
  2. Freeze scope + success criteria

    • Restate the mission, constraints, and “done” criteria in concrete terms.
    • Identify any authoritative sources (docs/specs) and record what claims must be backed by evidence.
  3. Create a phase plan and keep it current

    • Use your environment’s planning mechanism (e.g., update_plan if available) to track phases and prevent drifting.
    • Prefer 4–7 steps; keep exactly one step in progress.
  4. Decompose into subsystems

    • Choose subsystems that can be audited independently (API surface, core logic, error handling, perf, integrations, tests, docs).
    • For each subsystem, define 2–5 invariants (what must always be true).
  5. Run dual independent audits per subsystem

    • Spawn two independent auditor sub-agents for each subsystem (auditA and auditB).
    • Tell them to work independently until reconciliation (no cross-talk).
    • Require evidence for every issue (repo location, deterministic repro, expected vs actual, severity).
  6. Reconcile audits into a single confirmed issue list

    • Compare auditA vs auditB outputs and keep only mutually confirmed issues (or independently verify disputed ones with explorer).
    • Track rejected candidates with a brief reason (weak evidence, out of scope, non-deterministic).
    • Use this reconciled list as the only input to implementation.
    • Reconciliation output:
      • Confirmed issues (only mutual)
      • Rejected candidates (reason)
      • Consensus achieved: YES/NO
  7. Implement in clusters with clear ownership

    • Group confirmed issues into clusters that can be fixed with minimal coupling.
    • Spawn exactly one implementer tier per cluster.
    • Assign each implementer a file set to “own” and require them to avoid broad refactors.
    • Do not implement any cluster work directly; always delegate to the implementer (even for “quick” changes).
    • Every fix must come with a regression test (unit/integration/e2e as appropriate).
    • For each cluster, run a two-stage review loop:
      • Have the implementer complete the cluster (tests, self-review) and report what changed.
      • spec_reviewer validates “nothing more, nothing less” by reading code (do not trust the report).
      • quality_reviewer validates maintainability and test quality (only after spec compliance passes).
      • If any review FAILs, send concrete feedback to the implementer and repeat the failed review stage.
  8. Enforce review gates

    • Do not merge/land a cluster unless spec compliance PASS and code quality PASS are both recorded with concrete references.
  9. Integrate + validate

    • Run the repo’s standard validations (tests, lint, build, typecheck).
    • If the repo has no clear commands, discover them from README, package.json, pyproject.toml, CI config, etc.
  10. Deliver a concise completion report

    • State what is usable now.
    • State what remains intentionally unsupported (with next steps/issues).
    • List commands executed (at least key validation commands) and results.

What to send to sub-agents

Keep your messages task-specific and concise.

For any audit/review/implementation message, include:

  • Goal + success criteria (what “done” means)
  • Scope boundaries / owned files (what to touch, what not to touch)
  • Invariants (2–5) that must hold
  • Commands to run (if known), and what evidence to collect