Reviews investment banking pitch decks for number consistency across slides, data-narrative alignment, IB-standard language polish, and visual/formatting QC.
From financial-analysisnpx claudepluginhub anthropics/financial-services-plugins --plugin financial-analysisThis skill uses the workspace's default tool permissions.
references/ib-terminology.mdreferences/report-format.mdscripts/extract_numbers.pyDesigns and optimizes AI agent action spaces, tool definitions, observation formats, error recovery, and context for higher task completion rates.
Enables AI agents to execute x402 payments with per-task budgets, spending controls, and non-custodial wallets via MCP tools. Use when agents pay for APIs, services, or other agents.
Compares coding agents like Claude Code and Aider on custom YAML-defined codebase tasks using git worktrees, measuring pass rate, cost, time, and consistency.
Perform comprehensive QC on the presentation across four dimensions. Read every slide, then report findings.
This skill works in both the PowerPoint add-in and chat. Identify which you're in before starting:
.pptx file.This is read-and-report only — no edits — so the workflow is identical in both.
Pull text from every slide, keeping track of which slide each line came from. You'll need slide-level attribution for every finding ("$500M appears on slides 3 and 8, but slide 15 shows $485M"). A deck with 30 slides is too much to hold in working memory reliably — write the extracted text to a file so the number-checking script can process it.
The script expects markdown-ish input with slide markers. Format as:
## Slide 1
[slide 1 text content]
## Slide 2
[slide 2 text content]
Run the extraction script on what you collected:
python scripts/extract_numbers.py /tmp/deck_content.md --check
It normalizes units ($500M vs $500MM vs $500,000,000 → same number), categorizes values (revenue, EBITDA, multiples, margins), and flags when the same metric category shows conflicting values on different slides. This is the part most likely to catch something a human missed on the fifth read-through.
Beyond what the script flags, verify:
Map claims to the data that's supposed to support them. This is where decks go wrong quietly — someone edits the chart on slide 7 and forgets the narrative on slide 4.
IB decks have a register. Scan for anything that breaks it: casual phrasing ("pretty good", "a lot of"), contractions, exclamation points, vague quantifiers without numbers, inconsistent terminology for the same concept.
See references/ib-terminology.md for replacement patterns.
Run standard visual verification checks on each slide. You're looking for: missing chart source citations, missing axis labels, typography inconsistencies, number formatting drift (1,000 vs 1K within the same deck), date format drift, footnote and disclaimer gaps.
Visual verification catches overlaps, overflow, and contrast issues that don't show up in text extraction. Don't skip it — a chart with no source citation looks the same as a properly sourced one in the text dump.
Use references/report-format.md as the structure. Categorize by severity:
Lead with criticals. If there aren't any, say so explicitly — "no number inconsistencies found" is a finding, not an absence of one.