Generate or improve a company-specific data analysis skill by extracting tribal knowledge from analysts. BOOTSTRAP MODE - Triggers: "Create a data context skill", "Set up data analysis for our warehouse", "Help me create a skill for our database", "Generate a data skill for [company]" → Discovers schemas, asks key questions, generates initial skill with reference files ITERATION MODE - Triggers: "Add context about [domain]", "The skill needs more info about [topic]", "Update the data skill with [metrics/tables/terminology]", "Improve the [domain] reference" → Loads existing skill, asks targeted questions, appends/updates reference files Use when data analysts want Claude to understand their company's specific data warehouse, terminology, metrics definitions, and common query patterns.
Creates tailored data analysis skills by extracting company-specific data knowledge from analysts and their warehouse.
/plugin marketplace add sksdesignnew/claudepg/plugin install data@knowledge-work-pluginsThis skill inherits all available tools. When active, it can use any tool Claude has access to.
references/domain-template.mdreferences/example-output.mdreferences/skill-template.mdreferences/sql-dialects.mdscripts/package_data_skill.pyA meta-skill that extracts company-specific data knowledge from analysts and generates tailored data analysis skills.
This skill has two modes:
Use when: User wants to create a new data context skill for their warehouse.
Step 1: Identify the database type
Ask: "What data warehouse are you using?"
Common options:
Use ~~data warehouse tools (query and schema) to connect. If unclear, check available MCP tools in the current session.
Step 2: Explore the schema
Use ~~data warehouse schema tools to:
Sample exploration queries by dialect:
-- BigQuery: List datasets
SELECT schema_name FROM INFORMATION_SCHEMA.SCHEMATA
-- BigQuery: List tables in a dataset
SELECT table_name FROM `project.dataset.INFORMATION_SCHEMA.TABLES`
-- Snowflake: List schemas
SHOW SCHEMAS IN DATABASE my_database
-- Snowflake: List tables
SHOW TABLES IN SCHEMA my_schema
After schema discovery, ask these questions conversationally (not all at once):
Entity Disambiguation (Critical)
"When people here say 'user' or 'customer', what exactly do they mean? Are there different types?"
Listen for:
Primary Identifiers
"What's the main identifier for a [customer/user/account]? Are there multiple IDs for the same entity?"
Listen for:
Key Metrics
"What are the 2-3 metrics people ask about most? How is each one calculated?"
Listen for:
Data Hygiene
"What should ALWAYS be filtered out of queries? (test data, fraud, internal users, etc.)"
Listen for:
Common Gotchas
"What mistakes do new analysts typically make with this data?"
Listen for:
Create a skill with this structure:
[company]-data-analyst/
├── SKILL.md
└── references/
├── entities.md # Entity definitions and relationships
├── metrics.md # KPI calculations
├── tables/ # One file per domain
│ ├── [domain1].md
│ └── [domain2].md
└── dashboards.json # Optional: existing dashboards catalog
SKILL.md Template: See references/skill-template.md
SQL Dialect Section: See references/sql-dialects.md and include the appropriate dialect notes.
Reference File Template: See references/domain-template.md
Use when: User has an existing skill but needs to add more context.
Ask user to upload their existing skill (zip or folder), or locate it if already in the session.
Read the current SKILL.md and reference files to understand what's already documented.
Ask: "What domain or topic needs more context? What queries are failing or producing wrong results?"
Common gaps:
For the identified domain:
Explore relevant tables: Use ~~data warehouse schema tools to find tables in that domain
Ask domain-specific questions:
Generate new reference file: Create references/[domain].md using the domain template
Each reference file should include:
Before delivering a generated skill, verify:
Use when working with Payload CMS projects (payload.config.ts, collections, fields, hooks, access control, Payload API). Use when debugging validation errors, security issues, relationship queries, transactions, or hook behavior.