verification-before-completion
1
总安装量
1
周安装量
#53176
全站排名
安装命令
npx skills add https://github.com/rileyhilliard/claude-essentials --skill verification-before-completion
Agent 安装分布
amp
1
opencode
1
kimi-cli
1
github-copilot
1
antigravity
1
Skill 文档
Verification Before Completion
Core Principle: No completion claims without fresh verification evidence.
The Verification Gate
BEFORE any claim of success, completion, or satisfaction:
Copy this checklist and track your progress:
Verification Checklist:
- [ ] IDENTIFY: What command proves this claim?
- [ ] RUN: Execute the verification command (fresh, complete)
- [ ] READ: Check full output, exit code, failure counts
- [ ] VERIFY: Does output confirm the claim?
- If NO â State actual status with evidence
- If YES â State claim WITH evidence
- IDENTIFY – What command proves this claim?
- RUN – Execute the full verification command (fresh, complete)
- READ – Check full output, exit code, failure counts
- VERIFY – Does output confirm the claim?
- NO â State actual status with evidence
- YES â State claim WITH evidence from step 2-3
Skip any step = invalid claim.
When This Applies
ALWAYS before:
- Claiming “tests pass”, “build succeeds”, “linter clean”, “bug fixed”
- Expressing satisfaction (“Great!”, “Done!”, “Perfect!”)
- Using qualifiers (“should work”, “probably fixed”, “seems to”)
- Committing, creating PRs, marking tasks complete
- Moving to next task or delegating work
- ANY statement implying success or completion
Common Verification Requirements
| Claim | Required Evidence | Not Sufficient |
|---|---|---|
| Tests pass | yarn test output: 0 failures |
Previous run, “looks correct” |
| Build succeeds | Build command: exit 0 | Linter clean, “should work” |
| Bug fixed | Test reproducing bug: now passes | Code changed, assumed fix |
| Linter clean | Linter output: 0 errors | Partial check, spot test |
| Regression test works | RedâGreen cycle verified | Test passes once |
| Agent task complete | VCS diff shows expected changes | Agent reports “success” |
Red Flags
Stop and verify if you’re about to:
- Use hedging language (“should”, “probably”, “seems to”)
- Express satisfaction before running verification
- Trust agent/tool success reports without independent verification
- Rely on partial checks or previous runs
- Think “just this once” or “I’m confident it works”
Key Examples
Regression test (TDD Red-Green):
â
Write test â Run (fail) â Fix code â Run (pass) â Revert fix â Run (MUST fail) â Restore â Run (pass)
â "I've written a regression test" (without verifying red-green cycle)
Build vs Linter:
â
Run `npm run build` â See "exit 0" â Claim "build passes"
â Run linter â Claim "build will pass" (linter â compiler)
Agent delegation:
â
Agent reports success â Check `git diff` â Verify changes â Report actual state
â Trust agent's success message without verification
Why It Matters
Unverified claims break trust and ship broken code:
- Undefined functions that crash production
- Incomplete features missing requirements
- Lost time on rework after false completion
- Partner distrust: “I don’t believe you”
Violating this skill violates core honesty requirements.
No Exceptions
Run the command. Read the output. THEN claim the result.
Evidence before assertions, always.