From togetherai-skills
Generates dense vector embeddings, enables semantic search, builds RAG pipelines, and reranks results via Together AI. Use for vector representations and retrieval improvements over direct generation.
npx claudepluginhub togethercomputer/skillsThis skill uses the workspace's default tool permissions.
Use this skill for semantic retrieval components:
Guides selection and optimization of embedding models for vector search and RAG, including chunking strategies, dimension reduction, multilingual support, with Python templates.
Selects and optimizes embedding models for RAG and semantic search. Guides chunking strategies, model comparisons, dimension reduction, multilingual support, and Python templates.
Build RAG systems for LLM apps using vector databases, embeddings, and retrieval strategies. Use for document Q&A, grounded chatbots, and semantic search.
Share bugs, ideas, or general feedback.
Use this skill for semantic retrieval components:
This skill is for retrieval plumbing, not for the final language-model response itself.
together-chat-completions for the final answer-generation steptogether-batch-inference for very large offline embedding backfillstogether-dedicated-endpoints when reranking requires a dedicated deploymentsemantic_search.py). Use a dedicated vector database for production scale.together>=2.0.0). If the user is on an older version, they must upgrade first: uv pip install --upgrade "together>=2.0.0".rag_pipeline.py example demonstrates retrieval plus generation; treat generation as a hand-off to chat completions.