From together-pack
Guides performance tuning for Together AI inference, fine-tuning, and model deployment using OpenAI-compatible API. Covers errors, models, batch inference, and resources.
npx claudepluginhub jeremylongshore/claude-code-plugins-plus-skills --plugin together-packThis skill is limited to using the following tools:
Guidance for performance tuning with Together AI inference and fine-tuning API.
Guides Together AI API integration for inference, fine-tuning, and model deployment using OpenAI-compatible clients and Python SDK.
Guides LoRA, full fine-tuning, DPO preference tuning, VLM training, function-calling tuning, and reasoning tuning on Together AI using custom datasets to adapt models.
Guides Next.js Cache Components and Partial Prerendering (PPR): 'use cache' directives, cacheLife(), cacheTag(), revalidateTag() for caching, invalidation, static/dynamic optimization. Auto-activates on cacheComponents: true.
Share bugs, ideas, or general feedback.
Guidance for performance tuning with Together AI inference and fine-tuning API.
base_url = 'https://api.together.xyz/v1'together Python SDK or any OpenAI client library| Error | Cause | Solution |
|---|---|---|
401 Unauthorized | Invalid API key | Check at api.together.xyz |
Model not found | Wrong model ID | Use client.models.list() |
429 Rate limit | Too many requests | Implement backoff |
500 Server error | Model overloaded | Retry with backoff |
See related Together AI skills for more patterns.