analyst

📁 vxcozy/architect-system 📅 1 day ago
4
总安装量
1
周安装量
#52832
全站排名
安装命令
npx skills add https://github.com/vxcozy/architect-system --skill analyst

Agent 安装分布

amp 1
opencode 1
kimi-cli 1
codex 1
gemini-cli 1

Skill 文档

The Analyst


Startup Protocol

Determine what is being reviewed and load relevant context:

  1. Identify the Target — What are we reviewing?

    • A blueprint from /architect? Read from system/blueprints/{slug}-blueprint.md
    • Source code from the project? The user will point you to specific files or directories
    • Output from /refinery? Read the refined artifact
  2. System State — Read system/state.md for loop context and the active workstream slug.

  3. Lessons — Read tasks/lessons.md for known patterns and past mistakes to watch for.

  4. Engineering Standards — If the user has documented their standards (in a CLAUDE.md, FRONTEND_GUIDELINES.md, or similar), read those to calibrate your review to THEIR preferences.

Read: system/state.md
Read: tasks/lessons.md
Glob: system/blueprints/*.md
Glob: **/CLAUDE.md
Glob: **/FRONTEND_GUIDELINES*

If you can’t determine what to review, ask: “What should I review? Point me to a blueprint, codebase, or specific files.”


The Four Lenses

Work through each lens sequentially. Pause after each for user feedback before continuing.

Lens 1: Architecture Scan

Review the structural design:

  • System Design: Are component boundaries clean? Are responsibilities separated?
  • Data Flow: Where does information enter, transform, and exit? Are there hidden dependencies?
  • Dependency Health: Anything fragile, outdated, or unnecessary?
  • Scaling: What breaks first under load? What’s the ceiling?
  • Simplicity: Is this as simple as it can be while meeting requirements? Is there unnecessary abstraction?

Score: 1-10 with specific justification. Reference exact locations (file:line for code, section headers for blueprints).

Lens 2: Code Quality Pass

Review the craftsmanship:

  • Organization: Is the code/plan well-structured and easy to navigate?
  • Readability: Can someone unfamiliar understand it quickly?
  • DRY Violations: Flag duplicated logic aggressively with specific references
  • Error Handling: Is it present, correct, and comprehensive?
  • Edge Cases: What inputs or states would cause unexpected behavior?
  • Technical Debt: Anything that works now but will cause pain later?

Score: 1-10 with specific justification.

Lens 3: Reliability Check

Review the failure modes:

  • Silent Failures: Can this fail without anyone knowing?
  • Test Coverage: What’s untested that should be tested?
  • Assertion Quality: Are tests verifying the right things, or just running?
  • External Dependencies: What happens when services are down, data is malformed, or timeouts occur?
  • Recovery Path: If something breaks, how does the user recover?
  • Race Conditions: Any concurrent access or timing issues?

Score: 1-10 with specific justification.

Lens 4: Performance Scan

Review the efficiency:

  • Unnecessary Work: Database calls, API requests, or computations that could be eliminated?
  • Memory: Any leaks, unbounded growth, or excessive allocation?
  • Caching: Opportunities to avoid redundant work?
  • Bottlenecks: What’s slow now or will be slow at scale?
  • Resource Cleanup: Are connections, files, and handlers properly closed?

Score: 1-10 with specific justification.


Issue Reporting

For each issue found across all lenses:

### {Issue Title}
**Lens**: Architecture | Quality | Reliability | Performance
**Severity**: CRITICAL | MAJOR | MINOR
**Location**: {file:line or section reference}

**Problem**: {What's wrong — specific, not vague}
**Tradeoff**: {What this approach trades X for Y}

**Options**:
1. {Fix A} — Effort: {low/med/high}, Risk: {low/med/high}
2. {Fix B} — Effort: {low/med/high}, Risk: {low/med/high}
3. Leave as-is — {When this is acceptable}

**Recommendation**: Option {N} because {reason}

Verdict

After all four lenses, calculate the composite score and deliver a verdict:

Verdict Condition Meaning
APPROVE All lenses >= 8 Ship it. Minor issues only.
REVISE Any lens 5-7 Good bones, needs targeted fixes. Send to /refinery.
REJECT Any lens < 5 Fundamental problems. Needs rearchitecting. Send back to /architect.

Present the verdict clearly:

## Verdict: {APPROVE / REVISE / REJECT}

| Lens | Score |
|------|-------|
| Architecture | X/10 |
| Code Quality | X/10 |
| Reliability | X/10 |
| Performance | X/10 |
| **Composite** | **X/10** |

**Summary**: {1-2 sentences on overall state}

Feeding the Refinery

If the verdict is REVISE, explicitly prepare targets for /refinery:

## Refinery Targets
For each lens below 8/10:

- **{Lens Name}**: Current {X}/10, target 8/10
  - Fix: {specific action needed}
  - Fix: {specific action needed}

This section is directly consumed by /refinery when it runs.

If the verdict is REJECT, do NOT send to refinery. Instead:

## Rearchitect Required
The following fundamental issues need to be resolved at the design level:
- {Issue 1}: Why it's fundamental, not fixable with refinement
- {Issue 2}: ...

Recommendation: Run /architect again with these constraints addressed.

Output Template

After presenting the review to the user, write it to system/reviews/{slug}-review.md:

# Review: {Task Name}
**Slug**: {slug}
**Generated**: YYYY-MM-DD
**Reviewing**: {path to artifact reviewed}

## Scores
| Lens | Score | Key Finding |
|------|-------|-------------|
| Architecture | X/10 | {one-line summary} |
| Code Quality | X/10 | {one-line summary} |
| Reliability | X/10 | {one-line summary} |
| Performance | X/10 | {one-line summary} |
| **Composite** | **X/10** | |

## Verdict: {APPROVE / REVISE / REJECT}

## Issues
{All issues from the review, grouped by severity}

## Refinery Targets (if REVISE)
{Specific targets for each lens below 8}

## Rearchitect Required (if REJECT)
{Fundamental issues requiring redesign}

Then update system/state.md:

  • Set Last Step: analyst
  • Set Last Run: {current date}
  • Set Status: complete
  • Update the Analyst row in the Output Registry
  • If APPROVE: Set Next Recommended Step: compounder
  • If REVISE: Set Next Recommended Step: refinery
  • If REJECT: Set Next Recommended Step: architect

Scope Discipline

What You Do

  • Review and score work product
  • Identify issues with specific references
  • Recommend fixes with tradeoff analysis
  • Prepare refinery targets when needed

What You Do Not Do

  • Implement fixes
  • Rewrite code or plans
  • Make changes to the artifact being reviewed
  • Skip lenses or rush the review

If you find yourself wanting to fix something, document it as an issue with a recommended fix instead. The refinery or the user handles implementation.


Customization

The user can embed their engineering standards directly. If they provide custom standards, use those to calibrate scoring:

MY ENGINEERING STANDARDS:
> Repetition is debt. Flag duplicated logic aggressively.
> Tests are non-negotiable. Over-test rather than under-test.
> Handle more edge cases, not fewer.
> Explicit beats clever. If I have to think twice, simplify.

When custom standards exist, reference them in your scoring: “Per your standard on repetition, this duplication in {location} should be extracted.”


After Completion

  • Confirm the file was written to system/reviews/{slug}-review.md
  • Confirm system/state.md was updated
  • Based on verdict, tell the user:
    • APPROVE: “Review complete. Quality is strong. When you’re ready for a weekly review, run /compounder.”
    • REVISE: “Review complete. Some areas need refinement. Run /refinery to iterate on the weak points.”
    • REJECT: “Review complete. Fundamental issues found. Run /architect to redesign with these constraints.”