From mlx
Search, fetch, download, and extract ML/AI research papers from 7 free academic sources. Find and download ML datasets from 5 free sources (HuggingFace, OpenML, UCI, Papers with Code, Kaggle). Review a paper, critique methodology, assess reproducibility, evaluate experimental design. Convert research papers, articles, or technical documents into working code prototypes. Use when the user wants to find papers, search arxiv, get citations, download a PDF, extract text from a paper, find/download datasets, review/critique a research paper, implement a paper, prototype an algorithm, or convert research to working code.
npx claudepluginhub damionrashford/mlx --plugin mlxThis skill is limited to using the following tools:
Instructions and tools for searching, fetching, and extracting ML/AI research papers, plus finding and downloading datasets.
evals/evals.jsonreferences/api-reference.mdreferences/prototype/analysis-methodology.mdreferences/prototype/extraction-patterns.mdreferences/prototype/generation-rules.mdreferences/sources.mdscripts/analyze_document.pyscripts/datasets.pyscripts/download.pyscripts/extract.pyscripts/fetch.pyscripts/prototype/__init__.pyscripts/prototype/analyzers/__init__.pyscripts/prototype/analyzers/code_detector.pyscripts/prototype/analyzers/content_analyzer.pyscripts/prototype/extractors/__init__.pyscripts/prototype/extractors/markdown_extractor.pyscripts/prototype/extractors/notebook_extractor.pyscripts/prototype/extractors/pdf_extractor.pyscripts/prototype/extractors/web_extractor.pyGenerates design tokens/docs from CSS/Tailwind/styled-components codebases, audits visual consistency across 10 dimensions, detects AI slop in UI.
Records polished WebM UI demo videos of web apps using Playwright with cursor overlay, natural pacing, and three-phase scripting. Activates for demo, walkthrough, screen recording, or tutorial requests.
Delivers idiomatic Kotlin patterns for null safety, immutability, sealed classes, coroutines, Flows, extensions, DSL builders, and Gradle DSL. Use when writing, reviewing, refactoring, or designing Kotlin code.
Instructions and tools for searching, fetching, and extracting ML/AI research papers, plus finding and downloading datasets.
which pdftotext || echo "MISSING: brew install poppler (macOS) or apt install poppler-utils (Linux)"
| Source | Search | Fetch | Best for |
|---|---|---|---|
| arXiv | yes | yes | ML/AI preprints |
| Semantic Scholar | yes | yes | Citations, open-access PDFs |
| Papers with Code | yes | yes | Papers linked to GitHub repos |
| Hugging Face | yes | via arXiv | Trending daily papers |
| JMLR | yes | yes | Peer-reviewed ML journal |
| ACL Anthology | no | by ID | NLP conference papers |
| OpenScholar | no | no | Q&A synthesis over 45M papers (URL only) |
uv run ${CLAUDE_SKILL_DIR}/scripts/search.py "$ARGUMENTS" --source arxiv --max 5
Flags: --source (arxiv, semantic_scholar, papers_with_code, huggingface, jmlr, openscholar), --max N, --cat cs.AI,cs.LG, --sort relevance|date
uv run ${CLAUDE_SKILL_DIR}/scripts/fetch.py <paper_id>
Auto-detects source: 2401.12345 (arXiv), 2022.acl-long.220 (ACL), v22/19-920 (JMLR), 40-char hex (Semantic Scholar)
uv run ${CLAUDE_SKILL_DIR}/scripts/download.py <paper_id> -o ./papers/
uv run ${CLAUDE_SKILL_DIR}/scripts/extract.py ./papers/<file>.pdf --max-pages 20
--max-pages 10 for summaries; go deeper only when asked.uv run ${CLAUDE_SKILL_DIR}/scripts/scientific_search.py "$ARGUMENTS" --max 10
Searches arXiv + Semantic Scholar concurrently with deduplication. Add --datasets for Kaggle/HuggingFace datasets.
uv run ${CLAUDE_SKILL_DIR}/scripts/analyze_document.py <url_or_path> [--max-pages 10] [--json]
Extracts text from PDFs, Word docs, text files. Supports URLs and local paths.
| Source | Search | Info | Download | Best for |
|---|---|---|---|---|
| HuggingFace | yes | yes | yes (parquet) | NLP, vision, audio datasets |
| OpenML | yes | yes | yes (ARFF/CSV) | Tabular/benchmark datasets |
| UCI | yes | yes | yes (CSV/ZIP) | Classic ML datasets |
| Papers with Code | yes | yes | no (links only) | Datasets linked to papers |
| Kaggle | yes | no | no (use kaggle CLI) | Competition & community datasets |
uv run ${CLAUDE_SKILL_DIR}/scripts/datasets.py search "$ARGUMENTS" --source huggingface --limit 10
Flags: --source (huggingface, openml, uci, paperswithcode, kaggle), --limit N
uv run ${CLAUDE_SKILL_DIR}/scripts/datasets.py info <dataset_id> --source huggingface
Shows description, columns, splits, download URLs. Works with: huggingface, openml, uci, paperswithcode.
uv run ${CLAUDE_SKILL_DIR}/scripts/datasets.py download <dataset_id> --source huggingface --output ./datasets --split train
Downloads to ./datasets/ directory. Flags: --output DIR, --split train|test|validation (HuggingFace only).
info to check columns/splits before downloading.pd.read_parquet()).kaggle datasets download -d <owner/dataset>arXiv: 3s, Semantic Scholar: 4s, Papers with Code: 3s, JMLR: 3s per volume, OpenML: 2s, UCI: 2s, Kaggle: 2s.
Structured framework for reviewing ML/AI research papers. Produces fair, constructive, conference-quality reviews.
Use the research scripts to get the paper content:
# Download by arXiv ID
uv run ${CLAUDE_SKILL_DIR}/scripts/download.py 2401.12345 --output ./papers
# Extract text from PDF
uv run ${CLAUDE_SKILL_DIR}/scripts/extract.py ./papers/2401.12345.pdf --max-pages 30
# Fetch metadata
uv run ${CLAUDE_SKILL_DIR}/scripts/fetch.py 2401.12345
If the user provides only a topic, search first, then review the selected paper.
2-3 sentences: What is the paper about? What is the key contribution?
Evaluate each dimension:
Identify specific issues:
| Dimension | Assessment | Notes |
|---|---|---|
| Splits | proper / questionable / missing | Train/val/test separation |
| Baselines | fair / unfair / missing | SOTA included? |
| Metrics | appropriate / limited / wrong | Multiple metrics? |
| Significance | reported / missing | Error bars, CIs, p-values |
| Ablations | thorough / partial / none | Component contributions |
3-5 specific questions that would strengthen the paper or clarify ambiguities.
| Rating | |
|---|---|
| Recommendation | Accept / Weak Accept / Borderline / Weak Reject / Reject |
| Confidence | High / Medium / Low |
| Impact | What would this enable if results hold? |
When reviewing multiple papers on the same topic:
| Dimension | Paper A | Paper B | Paper C |
|---|---|---|---|
| Method | |||
| Dataset | |||
| Best metric | |||
| Reproducibility | |||
| Novelty |
Rank by overall contribution, noting complementary strengths.
Convert a research paper, article, or technical document into a complete working code project.
uv run ${CLAUDE_SKILL_DIR}/scripts/prototype/main.py <source> -o ./prototype [-l python] [-v]
| Argument | Description |
|---|---|
<source> | PDF file, URL, .ipynb, .md, or .txt |
-o | Output directory (default: ./prototype) |
-l | Language: python, javascript, typescript, rust, go |
-v | Verbose output |
Every generated project includes: main implementation, dependency manifest, test file, README, .gitignore.
| Domain | Default |
|---|---|
| ML / data science | Python |
| Web / frontend | TypeScript |
| Systems / performance | Rust |
Explicit numpy/torch mention | Python |
Explicit react/vue mention | JavaScript |
except)See references/prototype/generation-rules.md for full generation rules.