Fine-tune language models via Tinker API by diagnosing training issues like slow steps, hanging sessions, errors, and renderer problems, plus conduct post-training research: replicate papers, run SFT/RL/DPO experiments, tune hyperparameters, monitor runs, and analyze logs.
npx claudepluginhub thinking-machines-lab/tinker-cookbook --plugin tinkerDiagnose training issues with Tinker — slow steps, hanging sessions, output mismatches, error messages, renderer problems, and deployment issues. Use this skill whenever a user reports that training is slow, steps take too long, sessions are hanging, model outputs differ between Tinker and external engines (vLLM, SGLang), they get a confusing error message, training quality is poor (high KL, bad outputs), or they suspect something is wrong. Also trigger when users ask "is this a Tinker issue or my issue?", "is Tinker down?", report unexpected wait times, see output quality regressions, get opaque errors, or want to profile/debug their training or deployment pipeline. This skill walks through systematic triage to determine root cause.
Conduct post-training research for LLMs using the Tinker API — replicate paper results, explore new training ideas, run and monitor experiments, and document findings. Use this skill whenever the user wants to do research, replicate experiments from a paper or repo, investigate training hypotheses, run experiment sweeps, explore post-training techniques (SFT, RL, DPO, distillation, etc.), set up training, write training code, choose a model, tune hyperparameters, manage checkpoints, export weights, or analyze training logs — even if they just say "try this idea" or "let's see what happens if...".
Specialized skills for LLM engineering tasks including model training, evaluation, fine-tuning, and deployment optimization.
Share bugs, ideas, or general feedback.
Fine-tune LLMs end-to-end: env setup, LoRA training (SFT/DPO/GRPO/vision), evaluation, and export. Works on NVIDIA GPUs and Apple Silicon.
Transfer learning adaptation
ML engineering plugin: Give your AI coding agent ML engineering superpowers.
Agent Skills for AI/ML tasks including dataset creation, model training, evaluation, and research paper publishing on Hugging Face Hub
Ultra-compressed communication mode. Cuts ~75% of tokens while keeping full technical accuracy by speaking like a caveman.