From text-corpus-analysis
Decide whether a corpus analysis task should use classical NLP, a local LLM, or a cloud LLM (OpenRouter) given corpus size, task complexity, and cost tolerance. Use first, before any other skill in this plugin, especially when the corpus is large (thousands+ of documents) or when an LLM pass could get expensive.
npx claudepluginhub danielrosehill/claude-code-plugins --plugin text-corpus-analysisThis skill uses the workspace's default tool permissions.
Decision helper. Picks the right execution lane for a text-analysis task.
Provides UI/UX resources: 50+ styles, color palettes, font pairings, guidelines, charts for web/mobile across React, Next.js, Vue, Svelte, Tailwind, React Native, Flutter. Aids planning, building, reviewing interfaces.
Fetches up-to-date documentation from Context7 for libraries and frameworks like React, Next.js, Prisma. Use for setup questions, API references, and code examples.
Analyzes multiple pages for keyword overlap, SEO cannibalization risks, and content duplication. Suggests differentiation, consolidation, and resolution strategies when reviewing similar content.
Share bugs, ideas, or general feedback.
Decision helper. Picks the right execution lane for a text-analysis task.
| Lane | Strengths | Weaknesses | Cost model |
|---|---|---|---|
| Classical NLP (spaCy, scikit-learn, gensim, TextBlob, BlackLab) | Deterministic, fast, free, scales to millions of docs, strong for frequency/NER/TF-IDF/LDA | Brittle on short or noisy text; no semantic judgment | CPU time only |
| Local LLM (Ollama: llama3.1, qwen2.5, gemma2) | No per-token cost, private, good for classification/labeling on small batches | Slow on big corpora (GPU-bound), lower ceiling than frontier models | Electricity + wall-clock time |
| Cloud LLM (OpenRouter → Claude, GPT, Gemini, DeepSeek, Llama) | Best judgment, handles nuance, parallelizable | Per-token cost — dangerous on 10k+ docs without planning | $ per M tokens |
words × 1.3), compute total input tokens.Task: <restatement>
Corpus: N docs, ~M tokens total
Recommended lane: <Classical NLP | Local LLM | Cloud LLM via OpenRouter>
Recommended model (if LLM): <name> — est. $X.XX for full corpus
Cheaper fallback: <sampling strategy OR cheaper model> — est. $Y.YY
Reasoning: <2-3 sentences>