From armory
Audits academic or technical manuscripts for macro-coherence, argumentative structure, citation hygiene, visual formatting, and submission readiness, delivering a prioritized section-level refactoring report.
npx claudepluginhub mathews-tom/armory --plugin armoryThis skill uses the workspace's default tool permissions.
Execute a comprehensive, multi-pass diagnostic audit of an academic or
Implements Playwright E2E testing patterns: Page Object Model, test organization, configuration, reporters, artifacts, and CI/CD integration for stable suites.
Guides Next.js 16+ Turbopack for faster dev via incremental bundling, FS caching, and HMR; covers webpack comparison, bundle analysis, and production builds.
Discovers and evaluates Laravel packages via LaraPlugins.io MCP. Searches by keyword/feature, filters by health score, Laravel/PHP compatibility; fetches details, metrics, and version history.
Execute a comprehensive, multi-pass diagnostic audit of an academic or technical manuscript, producing a structured improvement report that identifies issues across 24 audit dimensions — from macro-coherence and argumentative architecture through claims-evidence calibration, narrative flow, prose microstructure, rendered visual inspection, and cross-element coherence, down to citation hygiene and reproducibility.
The output is a prioritized, actionable improvement plan — not a line edit. The goal is to surface structural, logical, and clarity issues that authors systematically miss because they're too close to the text.
Optimized for arXiv/preprint submissions with flexible compliance standards.
Companion skill: manuscript-provenance audits whether manuscript content
(numbers, tables, figures, ordering, terminology) is computationally derived
from code and scripts. This skill audits the document as prose; that skill
audits computational grounding. Run both for complete pre-publication coverage.
| Concern | This skill (manuscript-review) | manuscript-provenance |
|---|---|---|
| Reproducibility | Does the paper describe enough to reproduce? (§6) | Does the code actually produce what the paper claims? (§1, §7) |
| Figures/Tables | Legible, accessible, well-formatted? (§12) | Generated by scripts, not manual entry? (§2, §3) |
| Rendered visuals | Readable at print scale? Floats near references? (§23) | Figure generation script produces correct format? (§3) |
| Hyperparameters | Listed in the paper with rationale? (§6) | Values trace to config files, not hardcoded? (§1, §8) |
| Code availability | Statement exists in the paper? (§17) | Repo URL valid, README accurate, pipeline works? (§11) |
| Terminology | Abbreviations consistent within document? (§14) | Terms match code identifiers? (§5) |
| Significant figures | Consistent precision within document? (§12) | Precision matches script output? (§2) |
| Figure format | Appropriate format for document quality? (§12) | Format generated by script, not manually exported? (§3) |
| Computational cost | Reported in the paper? (§7) | Values trace to benchmarking scripts? (§1) |
| Macro-prose coherence | Prose framing appropriate for injected value? (§24) | Value traced to code, macro manifest produced? (§4) |
| Cross-element consistency | Prose, captions, figures, tables mutually consistent? (§24) | All elements from same run/pipeline output? (§9) |
Rule: This skill never opens the codebase. manuscript-provenance never judges prose quality. Each reads the other's report when available.
Integration point — Macro Manifest: manuscript-provenance produces a
macro manifest as part of its §4 audit: a structured list of every
macro-injected value, its resolved numeric value, its source (script + output
file), and its location(s) in the manuscript text. This skill's Pass 13
(Cross-Element Coherence) consumes that manifest to check whether the prose
surrounding each injected value is appropriate for the actual value. If no
provenance report exists, this skill extracts macro values directly from
.tex source (less precise — no source tracing, but coherence check still
runs).
Read the uploaded manuscript. Accept PDF, DOCX, LaTeX source, or Markdown. If multiple files are uploaded (e.g., main text + supplementary), process all of them.
Identify:
For arXiv submissions, compliance checks are advisory. Focus on technical quality, reproducibility, and clarity rather than strict formatting rules.
Read references/checklist.md — the comprehensive 24-section, ~175-checkpoint
refactoring checklist. Every audit pass is structured against this checklist.
Read references/checklist.md
Execute the following passes sequentially. Each pass maps to one or more checklist sections. Work systematically — for each checkpoint:
Pass 1 — Structural Integrity (Checklist §1, §4, §5, §10)
Pass 2 — Abstract & Title Calibration (Checklist §2, §3)
Pass 3 — Technical Rigor (Checklist §6, §7)
Pass 4 — Argumentation Quality (Checklist §8, §9)
Pass 5 — Citation & Reference Hygiene (Checklist §11)
Pass 6 — Visual & Tabular Quality (Checklist §12)
Pass 7 — Prose Mechanics (Checklist §13, §14, §15)
Pass 7b — AI-Pattern Detection (advisory)
Scan prose sections for residual AI-writing patterns using detection rules
from references/detection-patterns.md. Academic manuscripts
drafted or polished with AI assistants often retain detectable tells.
Focus on patterns relevant to academic writing:
Skip patterns that are acceptable in academic prose:
This pass is MEDIUM priority. Flag findings but do not over-correct — academic conventions overlap with some AI patterns. Severity: report individual instances as LOW, flag clusters of 3+ patterns in a single paragraph as MEDIUM.
Pass 8 — Best Practices & Reproducibility (Checklist §16, §17, §18, §19)
Pass 9 — Claims-Evidence Calibration (Checklist §20)
This is a dedicated pass through every assertion in the manuscript.
For each claim:
This pass is HIGH priority. Claims-evidence mismatch is the single most common reason reviewers reject papers. An overclaim in the abstract poisons the entire reading.
Pass 10 — Narrative Flow & Coherence (Checklist §21)
Read the manuscript linearly, tracking the reader's cognitive state. At each sentence and paragraph boundary, check:
Flag any location where a domain-expert reader would need to re-read, scroll back, or pause to reconstruct the logical connection. These are flow breaks.
This pass is HIGH priority. Papers with strong results but poor narrative flow exhaust reviewers. A reader who has to fight the text stops trusting the author.
Pass 11 — Prose Microstructure (Checklist §22)
Sentence-level and paragraph-level patterns that compound into readability problems:
This pass is MEDIUM priority on individual items but compounds — a manuscript with 20 ambiguous pronouns, 10 density spikes, and 5 dangling modifiers is materially harder to read even though no single instance is fatal.
Pass 12 — Rendered Document Inspection (Checklist §23)
This pass requires the compiled PDF. If only LaTeX source is provided, ask the user for the compiled PDF or compile it.
Open the PDF and inspect every page at actual print scale:
This pass is HIGH priority. A paper with illegible axis labels or a table split across pages signals carelessness to reviewers regardless of technical quality. These defects are invisible from source and the author often doesn't notice because they read the paper in their editor, not in the compiled output.
Pass 13 — Cross-Element Coherence (Checklist §24)
Read the manuscript as an integrated system. For each figure, table, and macro-injected value:
\ref points to the element the
surrounding prose describes. After figure reordering, references often
point to the wrong visual.If a manuscript-provenance report exists, load its macro manifest (list of
all traced macro values with locations and source values) and use it as
input for step 4. If no provenance report exists, extract macro values
directly from .tex source.
This pass is HIGH priority. Cross-element incoherence is the most insidious class of manuscript defect — each piece looks fine in isolation, the system is broken. Reviewers notice because they read the document linearly and encounter contradictions the author can't see because they edit pieces independently.
Note for arXiv: Ethics statements, anonymization, page limits, and strict formatting requirements are marked N/A by default. Focus on technical quality, reproducibility, and clarity.
Produce the report as a structured document. Use references/report-template.md
as the output format.
Read references/report-template.md
Report structure:
Executive Summary — Overall quality assessment (Publication-ready / Recommend revisions / Needs work). Top 5 high-priority improvements.
Per-Section Diagnostics — For each manuscript section, the specific issues found, mapped to checklist checkpoint IDs. Severity tagged as HIGH (impacts clarity/credibility), MEDIUM (noticeable quality gap), or LOW (polish/optional improvement).
Cross-Cutting Issues — Problems that span multiple sections (e.g., inconsistent notation, citation patterns, clarity patterns).
Priority Queue — All issues ranked by impact × effort. HIGH-impact items first, then MEDIUM items ordered by estimated fix effort (lowest effort first = quick wins).
Checklist Status — The full 24-section checklist with pass/needs-work/not-applicable status per checkpoint, referencing specific locations in the manuscript.
After completing the full scan, categorize issues:
For arXiv submissions, focus HIGH priority on technical quality and reproducibility. Compliance items (ethics statements, formatting) are typically LOW priority or N/A.
Present the priority queue first, then the detailed findings.
Save the report as a Markdown file in the same directory as the manuscript,
named [manuscript-name]-review-report.md.
Present the file to the user with a concise summary:
Focus on structure and clarity. This is a structural and technical audit. Sentence-level grammar is out of scope unless it forms a systematic pattern affecting readability.
Evidence-based findings. Every issue cites the specific manuscript location (section, paragraph, figure/table number). No vague "could be better."
Balanced severity. HIGH priority for technical credibility and reproducibility issues. MEDIUM for clarity and professional quality. LOW for style preferences. ArXiv allows more flexibility than peer-reviewed venues.
Context-aware recommendations. Formatting and compliance requirements vary by venue. For arXiv, prioritize technical quality over strict formatting. For journal submissions, adjust accordingly.
Constructive framing. Frame findings as improvements to clarity, credibility, and reproducibility rather than as rejection risks. ArXiv is more forgiving; focus on making the work accessible and trustworthy.
Direct communication. Report issues as issues with specific fixes, not as vague suggestions. But recognize that many "rules" are guidelines for arXiv.
Systematic coverage. Work through the checklist methodically. Mark items as pass/needs-work/N/A based on actual content. ArXiv-specific items (anonymization, page limits, strict templates) default to N/A.
User says any of:
All trigger this skill. Partial reviews (e.g., "just check citations") still run the full audit — the user benefits from comprehensive diagnostics even when they only asked about one aspect.