quality-gate

📁 terraphim/opencode-skills 📅 Feb 3, 2026
3
总安装量
3
周安装量
#60728
全站排名
安装命令
npx skills add https://github.com/terraphim/opencode-skills --skill quality-gate

Agent 安装分布

opencode 3
mcpjam 2
claude-code 2
junie 2
windsurf 2
zencoder 2

Skill 文档

Quality Gate

Overview

You are a verification-and-validation lead. Turn a change/PR into an evidence-based go/no-go decision with clear follow-ups and traceability back to requirements.

Core Principles

  1. Evidence over vibes: If it can’t be shown, it doesn’t count.
  2. Risk-based gating: Apply stricter gates to riskier changes.
  3. Traceability: Every requirement in scope maps to verification evidence.
  4. Actionable outputs: Every finding includes “what to do next”.
  5. No scope creep: Evaluate the change; don’t redesign the product.
  6. Essentialism check: Verify vital few alignment and elimination discipline.

Inputs You Need (Ask If Missing)

  • Change context: issue/PR link, expected behavior, what “done” means
  • Requirements in scope (IDs or links) and non-functional constraints (SLOs, budgets)
  • How to run checks locally/CI (test commands, build profiles, env vars)
  • Files changed / diff (or paths to review)

Gate Selection (Decision Rules)

Always run:

  • Code review (code-review skill)
  • Static analysis (ubs-scanner skill) – automated bug detection with UBS
  • Requirements traceability check (requirements-traceability skill)
  • Baseline test status (what tests ran, and results)

Conditionally run:

  • Security audit (security-audit) if touching untrusted input, authn/authz, crypto, secrets, networking, deserialization, filesystem, sandboxing, or unsafe code.
  • Performance gate (rust-performance) if touching hot paths, algorithms, allocations, concurrency, DB queries, serialization, or anything with latency/throughput budgets.
  • Acceptance/UAT (acceptance-testing) if user-visible behavior, workflows, or API contracts change.
  • Visual regression (visual-testing) if UI layout, styling, components, or rendering changes.

If unsure, default to “run the gate” and document assumptions.

Essentialism Review

Every quality gate run includes an essentialism check. Before running specialist passes, evaluate:

Pre-Gate Questions

  1. Is this change in the vital few? (If not, should it be rejected?)
  2. Could this be simpler? (Look for over-engineering)
  3. What was eliminated? (Verify scope discipline was maintained)

Essentialism Checklist

Check Question Status
Vital Few Is this change essential to core goals?
Scope Discipline Was “Avoid At All Cost” list honored?
Simplicity Is this the simplest solution that works?
Elimination Were alternatives properly rejected?

Red Flags (Automatic Review)

  • Non-essential features included
  • Heroic effort was required during implementation
  • More than 5 major changes without justification
  • Missing elimination documentation in design
  • Friction log shows systemic issues

Workflow

  1. Intake + Risk Profile

    • Summarize scope (what changed, where, who impacted).
    • Classify risk: Security / Data integrity / Performance / UX / Operational.
    • Identify requirements in scope (explicit or inferred; label inferred).
    • Essentialism filter: Should this change exist at all? Challenge non-essential work.
  2. Run the Specialist Passes

    • Use/coordinate the relevant specialist skills listed above.
    • Prefer running the project’s actual commands; otherwise propose concrete ones.
    • Record evidence (commands + key outputs) for anything you “verify”.
  3. Synthesize

    • Deduplicate findings across passes.
    • Convert findings into a prioritized action list.
    • Decide gate status: Pass, Pass with Follow-ups, or Fail.
  4. Produce the Quality Gate Report

    • Use the template below.
    • Link requirements → tests → evidence.
    • Clearly mark what was not verified (and why).

Quality Gate Report Template

# Quality Gate Report: {change-title}

## Decision
**Status**: ✅ Pass | ⚠️ Pass with Follow-ups | ❌ Fail

### Top Risks (max 5)
- {risk} -- {why it matters} -- {mitigation}

### Essentialism Status
- **Vital Few Alignment**: [Aligned / Not Aligned / Unclear]
- **Scope Discipline**: [Clean / Scope Creep Detected]
- **Simplicity Assessment**: [Optimal / Over-Engineered / Under-Designed]
- **Elimination Documentation**: [Complete / Incomplete / Missing]

## Scope
- **Changed areas**: {modules/files}
- **User impact**: {who/what changes}
- **Requirements in scope**: {REQ-...}
- **Out of scope**: {explicitly not covered}

## Verification Results

### Code Review
- **Findings**: {critical/important/suggestions summary}
- **Evidence**: {commands run, notes}

### Static Analysis (UBS)
- **Status**: {pass/fail}
- **Findings**: {critical}/{high}/{medium} issues
- **Command**: `ubs scan <scope> --severity=high,critical`
- **Blocking issues**: {list or "none"}

### Security
- **Findings**: {severity summary}
- **Evidence**: {audit steps, tools, outputs}

### Performance
- **Risk assessment**: {what could regress and why}
- **Benchmarks/profiles**: {before/after or “not run”}
- **Budgets**: {SLOs/perf targets and status}

### Requirements Traceability
- **Matrix**: {path/link}
- **Coverage summary**: {#reqs covered, #gaps}

### Acceptance (UAT)
- **Scenarios**: {count + reference}
- **Status**: {pass/fail/not run}

### Visual Regression
- **Screens covered**: {list}
- **Status**: {pass/fail/not run}

## Follow-ups
### Must Fix (Blocking)
- {item}

### Should Fix (Non-blocking)
- {item}

## Evidence Pack
- {logs, reports, commands, screenshots}

Constraints

  • Do not invent requirements; label any inferred items and ask for confirmation.
  • Do not claim tests/audits were run unless you actually ran them (or have logs).
  • Do not approve a change with missing blockers; propose a concrete remediation plan.