claims-extractor
15
总安装量
13
周安装量
#22123
全站排名
安装命令
npx skills add https://github.com/willoscar/research-units-pipeline-skills --skill claims-extractor
Agent 安装分布
claude-code
10
gemini-cli
9
codex
7
cursor
7
antigravity
6
Skill 文档
Claims Extractor (peer review)
Goal: turn a manuscript into an auditable list of claims that downstream skills can check.
Inputs
Required:
output/PAPER.md(or equivalent plain-text manuscript)
Optional:
DECISIONS.md(review scope or constraints)
Outputs
output/CLAIMS.md
Output format (recommended)
For each claim, include at minimum:
Claim: one sentenceType:empirical|conceptualScope: what the claim applies to / what it does not apply toSource: a locatable pointer intooutput/PAPER.md(section + page/figure/table + a short quote)
Workflow
- If
DECISIONS.mdexists, apply any review scope/format constraints. - Read the manuscript (
output/PAPER.md) end-to-end (at least abstract + intro + method + experiments + limitations). - Extract:
- primary contributions (what is new)
- key claims (what is asserted)
- assumptions (what must be true for claims to hold)
- Normalize each item into one sentence.
- Attach a source pointer for every item.
- Split into two sections:
- Empirical claims (must be backed by experiments/data)
- Conceptual claims (must be backed by argument/definition)
Definition of Done
-
output/CLAIMS.mdexists. - Every claim has a source pointer that can be located in
output/PAPER.md. - Empirical vs conceptual claims are clearly separated.
Troubleshooting
Issue: the paper is only a PDF or HTML
Fix:
- Convert/extract it into a plain-text
output/PAPER.mdfirst (even rough extraction is OK), then run claim extraction.
Issue: claims are vague (âsignificantâ, âbetterâ, âstate-of-the-artâ)
Fix:
- Rewrite each claim to include the measurable dimension (metric/dataset/baseline) or mark it as âunderspecifiedâ with a note.