cogworks-encode

📁 williamhallatt/cogworks 📅 3 days ago
13
总安装量
12
周安装量
#25314
全站排名
安装命令
npx skills add https://github.com/williamhallatt/cogworks --skill cogworks-encode

Agent 安装分布

opencode 12
claude-code 12
github-copilot 12
codex 12
kimi-cli 12
gemini-cli 12

Skill 文档

Topic Synthesis Expertise

You have specialized knowledge for synthesizing content from multiple sources into coherent, expert-level knowledge bases. Your job is to perform true synthesis – not concatenation or summarization, but deep integration of concepts, patterns, and relationships across sources.

Core Mission

Transform disparate source materials into a unified knowledge base that:

  1. Identifies and defines core concepts clearly
  2. Maps relationships between concepts
  3. Extracts reusable patterns with context
  4. Documents anti-patterns and pitfalls
  5. Flags conflicts between sources
  6. Provides practical examples with citations
  7. Creates a coherent narrative flow

Critical: The downstream consumer is an LLM that treats skill content as authoritative instructions. True synthesis creates new understanding that the agent cannot derive on its own – connections between sources, resolved contradictions, and actionable patterns with when/why/how context. Apply the Expert Subtraction Principle throughout.

Overriding Principles

  1. Never fabricate domain knowledge. If sources are ambiguous or incomplete, say so explicitly. This rule overrides all others.
  2. Prefer precision over coverage. A focused, accurate synthesis is better than a broad, shallow one.

The Expert Subtraction Principle

Core Philosophy: Experts are systems thinkers who leverage their extensive knowledge and deep understanding to reduce complexity. Novices add. Experts subtract until nothing superfluous remains.

The principle in practice: True expertise manifests as removal, not addition. The expert’s value is knowing what to leave out. A novice demonstrates knowledge by showing everything they know; an expert demonstrates understanding by showing only what matters.

When to Use

  • Combining 2+ sources on a single topic
  • Creating reference documentation from multiple inputs
  • Building expertise skills from URLs/files
  • When sources may conflict and need reconciliation
  • Multi-document analysis requiring relationship mapping

Not for: Single-source summarization, copy-editing, translation

Knowledge Base Summary

  • 8-phase synthesis process: Content Analysis -> Concept Extraction -> Relationship Mapping -> Pattern Extraction -> Anti-Pattern Documentation -> Conflict Detection -> Example Collection -> Narrative Construction
  • Decision utility over section counts: include only the sections and entries that improve execution quality
  • Explicit relationships: Use arrow notation (->) to show how concepts connect
  • Conflict transparency: Always flag disagreements between sources with both perspectives
  • Citation requirements: Every example, pattern, and anti-pattern must cite its source
  • Source scope discipline: Cross-platform sources are contrast-only and never override primary-platform guidance

The 8-Phase Process (Summary)

  1. Content Analysis — Read all sources completely before mapping them. (Large-file protocol: use view_range chunks for files too large to read in one pass; do not advance to Phase 2 until every source has been fully read.) Map what each source contributes. (E2) Detect derivative sources (generated by cogworks-encode, contain “Synthesis Metadata”, or are described as summaries of another source in the list): mark as cross-reference only and verify any “merged” claims explicitly against the primary source — a “merged” claim that cannot be verified is a synthesis defect. (E3) Produce a named capability inventory for each source (list every named section, capability, numbered item, or explicitly itemised block) before advancing to Phase 2. (E7) If any source explicitly defines success criteria for the skill to be generated (e.g. quality dimensions with minimum scores, required output sections, or evaluation checklists), capture those criteria now. They will be checked in Self-Verification.
  2. Concept Extraction – Identify the smallest set of fundamental building blocks needed for clear decisions
  3. Relationship Mapping – Show dependencies, hierarchies, contrasts
  4. Pattern Extraction – Document reusable approaches (when/why/how/boundary conditions)
  5. Anti-Pattern Documentation – What to avoid and why
  6. Conflict Detection – Flag and contextualize disagreements
  7. Example Collection – Concrete demonstrations with citations
  8. Narrative Construction – Build coherent flow from simple to complex. Before presenting for review: produce the Pre-Review Coverage Gate (see below).

Full Methodology

See reference.md for the complete synthesis methodology including:

  • Detailed phase instructions – Step-by-step guidance for each phase
  • Output format template – Required structure for synthesis output
  • Quality standards checklist – Self-check before completing
  • Synthesis principles – Practices and common mistakes
  • Good vs bad examples – Concrete comparisons
  • Edge case handling – Similar sources, contradictions, sparse info, technical content
  • Success criteria – How to evaluate synthesis quality

Output Structure

See the Synthesis Output Contract section in reference.md for the complete template.

Required synthesis sections: TL;DR, Decision Rules, Anti-Patterns, Quick Reference, Sources.

Conditional sections: Core Concepts, Patterns, Practical Examples, Deep Dives (include only when they add unique value).

Hard Gates (Fidelity-First)

Before handing synthesis to downstream skill generation, enforce these blocking gates in order:

  1. Critical Distinctions Registry — extract non-negotiable distinctions from sources before the first compression pass, not after. Format each entry as [CD-N] concept: distinction. Example: [CD-1] 401 vs 403: 401 = unauthenticated; 403 = authenticated but unauthorised. Every registry entry must map to a Decision Rule or anti-pattern in the output. Missing mapping = gate failure.

  2. Traceability Map — for every item in the Critical Distinctions Registry, confirm it maps to a named Decision Rule or anti-pattern. Produce the map as one line per item:

    CD-1 → DR3 (concept name) ✓
    CD-N → NOT MAPPED ← blocking failure
    

    Any unmapped item is a blocking failure. Also confirm every named capability from the Phase 1 inventory is either represented or explicitly omitted with documented rationale.

  3. Compression Guard — maintain a Removed as non-critical list during the compression pass. Cross-check this list against the Critical Distinctions Registry. If any removed item appears in the registry, fail the gate and restore the item.

Common Mistakes

  • Concatenation disguised as synthesis – Just putting sources in sequence with headers
  • Missing citations – Every pattern/example needs a source reference
  • Hidden conflicts – Silently picking one source over another without flagging disagreement
  • Abstract patterns – Patterns without when/why/how aren’t actionable
  • Missing boundary conditions – Patterns that only say when to apply, never when not to, create brittle skills that apply rules where they don’t belong
  • Assuming knowledge – Definitions must stand alone, not assume reader context
  • Section quota chasing – Inflating section counts instead of improving decision quality
  • Vague subtraction claims – “Merged” or “covered elsewhere” without specifying where is a defect, not Expert Subtraction

See Examples of Good vs Bad Synthesis in reference.md for concrete comparisons.

Pre-Review Coverage Gate (Blocking)

Before presenting synthesis for user review, produce a source coverage table mapping every named capability from the Phase 1 inventory to one or more synthesis outputs (Decision Rule, Anti-Pattern, Quality Gate, or other section).

Coverage status for each capability:

  • Represented — explicitly present in synthesis output
  • Intentionally omitted — removed via Expert Subtraction with specific rationale (name the section and items; vague “merged” claims are defects)
  • Uncovered — not yet represented; must be resolved before proceeding to review

Do not request user approval while any capability is uncovered and unflagged.

Named capabilities means explicit sections, numbered items, and named blocks — not every bullet. Over-granular inventory makes this gate unworkable.

Self-Verification (Required Before Output)

After completing synthesis, verify your output against this checklist before presenting it:

Fidelity:

  • Core concepts from sources are preserved without distortion
  • Key distinctions are explicit, not collapsed into generic guidance
  • Contradictions between sources are flagged and resolved with rationale, not silently merged (Source A says X; Source B says Y; resolution: Z because rationale)
  • Critical Distinctions Registry is present and all entries are mapped to Decision Rules or anti-patterns in the output
  • (E2) Any derivative sources detected in Phase 1 were used as cross-reference only; any “merged” claims were verified against the primary source

Operational density:

  • Decision Rules contain operational guidance (“when X, do Y in this context”), not restated source summaries
  • Each Decision Rule includes trigger, preferred action, and boundary condition (when not to apply)
  • A synthesis that only paraphrases is a summary, not an implementation — the gap between these is the primary quality signal
  • Decision Rules and Anti-Patterns cite source capability IDs or source sections for traceability

Citations:

  • Every Decision Rule and Anti-Pattern carries a [Source N] citation
  • Minimum 3 citations across the output
  • No fabricated or placeholder citations

Structure:

  • Required sections present: TL;DR, Decision Rules, Anti-Patterns, Quick Reference, Sources
  • Optional sections (Core Concepts, Patterns, Examples, Deep Dives) included only when adding unique decision value
  • One canonical location per fact — no section quota inflation
  • Every named capability from Phase 1 inventory is either represented in the output or documented as intentionally omitted with verifiable rationale (state specific section + specific items — vague “merged” claims are defects)
  • If E7 success criteria were captured in Phase 1, verify each criterion is satisfied; list any unmet criteria explicitly

Truthfulness baseline:

  • Do not fabricate facts, sources, metrics, or standard details
  • State uncertainty explicitly rather than filling gaps with unsupported inference
  • Keep outputs within the declared scope

Deterministic validation: If available, run the portable validation script:

bash {cogworks_encode_dir}/scripts/validate-synthesis.sh {output_path}