Help us improve
Share bugs, ideas, or general feedback.
From eval-runner
Publish an eval dataset to Hugging Face Hub (or GitHub as a fallback). Use when the user wants to share the inputs/labels used by an eval — with a dataset card, licensing, splits, and a content hash so downstream runs can verify integrity.
npx claudepluginhub danielrosehill/claude-eval-runner-pluginHow this skill is triggered — by the user, by Claude, or both
Slash command
/eval-runner:publish-datasetThe summary Claude sees in its skill listing — used to decide when to auto-load this skill
Packages a dataset from `datasets/<id>/` and pushes it to the Hub.
Mandates invoking relevant skills via tools before any response in coding sessions. Covers access, priorities, and adaptations for Claude Code, Copilot CLI, Gemini CLI.
Share bugs, ideas, or general feedback.
Packages a dataset from datasets/<id>/ and pushes it to the Hub.
datasets/.--target=<hf|github> — default hf.--private — private dataset repo.--license=<spdx> — default cc-by-4.0 (ask user if unclear).Pre-flight.
datasets/<id>/ exists with at least a data/ folder and a CARD.md draft. If no CARD.md, stub one and ask the user to fill load-bearing sections before publish.train/validation/test or the user's custom split.Normalise format. Prefer Parquet or JSONL with a consistent schema. Generate a schema.json. If source is CSV, convert. If source is images/audio, ensure a manifest file exists.
Hash. Compute a content hash (sha256 over sorted file hashes) and record it in datasets/<id>/HASH. This is what run-eval verifies.
Dataset card. Finish CARD.md with the HF standard sections: dataset summary, supported tasks, languages, dataset structure, data fields, data splits, source data, annotations, considerations for using the data, licensing, citation. Draft missing sections with TODO markers — don't fabricate.
Push.
hf: huggingface-cli repo create (dataset type), huggingface-cli upload. User defaults: the HF user is danielrosehill unless stated otherwise — confirm if ambiguous.github: create a repo, use Git LFS for anything > 50MB, push.Cross-link. Update the relevant eval's DATASET.md to point at the published URL and record the hash. If an eval config references the dataset, update it to pull from the Hub.
Record publication. Append to docs/publications.md.
Report. URL, hash, size, splits, license.