npx claudepluginhub jeremylongshore/claude-code-plugins-plus-skills --plugin groq-packThis skill is limited to using the following tools:
Set up a fast, reproducible local development workflow for Groq. Groq's sub-second response times make it uniquely suited for tight dev loops -- you get LLM responses fast enough to iterate without context-switching.
Applies production Groq SDK patterns in TypeScript and Python for typed clients, chat completions, usage tracking, and streaming.
Sets up Node.js/TypeScript dev environment for Mistral AI with tsx hot reload, Vitest unit/integration tests including mocking, and dotenv env config.
Sets up local dev workflow for Anthropic Claude API projects with singleton client, request logging, cost tracking, mocks, and hot reload.
Share bugs, ideas, or general feedback.
Set up a fast, reproducible local development workflow for Groq. Groq's sub-second response times make it uniquely suited for tight dev loops -- you get LLM responses fast enough to iterate without context-switching.
groq-sdk installedGROQ_API_KEY set (free tier is fine for development)my-groq-project/
├── src/
│ ├── groq/
│ │ ├── client.ts # Singleton Groq client
│ │ ├── models.ts # Model constants and selection
│ │ └── completions.ts # Completion wrappers
│ └── index.ts
├── tests/
│ ├── groq.test.ts # Unit tests with mocks
│ └── groq.integration.ts # Live API tests (CI-only)
├── .env.local # Local secrets (git-ignored)
├── .env.example # Template for team
└── package.json
{
"scripts": {
"dev": "tsx watch src/index.ts",
"test": "vitest",
"test:watch": "vitest --watch",
"test:integration": "GROQ_INTEGRATION=1 vitest tests/groq.integration.ts"
},
"dependencies": {
"groq-sdk": "^0.12.0"
},
"devDependencies": {
"tsx": "^4.0.0",
"vitest": "^2.0.0"
}
}
// src/groq/client.ts
import Groq from "groq-sdk";
let _client: Groq | null = null;
export function getGroqClient(): Groq {
if (!_client) {
if (!process.env.GROQ_API_KEY) {
throw new Error("GROQ_API_KEY not set. Copy .env.example to .env.local");
}
_client = new Groq({
apiKey: process.env.GROQ_API_KEY,
maxRetries: 2,
timeout: 30_000,
});
}
return _client;
}
// Reset for testing
export function resetClient(): void {
_client = null;
}
// src/groq/models.ts
export const MODELS = {
FAST: "llama-3.1-8b-instant", // Dev default: cheapest, fastest
VERSATILE: "llama-3.3-70b-versatile", // Production quality
SPECDEC: "llama-3.3-70b-specdec", // Speculative decoding variant
SCOUT: "meta-llama/llama-4-scout-17b-16e-instruct", // Vision
} as const;
export const DEV_MODEL = MODELS.FAST; // Use 8B for dev to save quota
// tests/groq.test.ts
import { describe, it, expect, vi, beforeEach } from "vitest";
import Groq from "groq-sdk";
// Mock the entire groq-sdk module
vi.mock("groq-sdk", () => {
const mockCreate = vi.fn().mockResolvedValue({
choices: [{ message: { content: "mocked response" }, finish_reason: "stop" }],
usage: { prompt_tokens: 10, completion_tokens: 5, total_tokens: 15 },
model: "llama-3.1-8b-instant",
});
return {
default: vi.fn(() => ({
chat: { completions: { create: mockCreate } },
models: { list: vi.fn().mockResolvedValue({ data: [] }) },
})),
};
});
describe("Groq Completions", () => {
it("should create a chat completion", async () => {
const groq = new Groq();
const result = await groq.chat.completions.create({
model: "llama-3.1-8b-instant",
messages: [{ role: "user", content: "test" }],
});
expect(result.choices[0].message.content).toBe("mocked response");
expect(result.usage.total_tokens).toBe(15);
});
});
// tests/groq.integration.ts
import { describe, it, expect } from "vitest";
import Groq from "groq-sdk";
const shouldRun = !!process.env.GROQ_INTEGRATION;
describe.skipIf(!shouldRun)("Groq Integration", () => {
const groq = new Groq();
it("should list available models", async () => {
const models = await groq.models.list();
expect(models.data.length).toBeGreaterThan(0);
const ids = models.data.map((m) => m.id);
expect(ids).toContain("llama-3.1-8b-instant");
}, 10_000);
it("should complete a chat request", async () => {
const result = await groq.chat.completions.create({
model: "llama-3.1-8b-instant",
messages: [{ role: "user", content: "Reply with exactly: PONG" }],
temperature: 0,
max_tokens: 10,
});
expect(result.choices[0].message.content).toContain("PONG");
}, 10_000);
});
# .env.example
# Get your key at https://console.groq.com/keys
GROQ_API_KEY=gsk_your_key_here
# Optional: override default dev model
GROQ_MODEL=llama-3.1-8b-instant
llama-3.1-8b-instant during development (lowest quota usage, fastest)temperature: 0 for deterministic outputs during debuggingmax_tokens conservatively to avoid burning through free tier| Error | Cause | Solution |
|---|---|---|
GROQ_API_KEY not set | Missing .env.local | Copy from .env.example |
| Test timeout | Live API call in unit test | Mock groq-sdk in unit tests |
429 rate_limit_exceeded | Free tier RPM hit | Wait 60s or use test:watch with longer intervals |
| Port already in use | Another tsx watch running | Kill process or change port |
See groq-sdk-patterns for production-ready code patterns.