From langchain-pack
Configures GitHub Actions CI/CD for LangChain apps: mocked Vitest unit tests, gated real-LLM integration tests, RAG validation, LangSmith tracing, and TypeScript checks.
npx claudepluginhub jeremylongshore/claude-code-plugins-plus-skills --plugin langchain-packThis skill is limited to using the following tools:
CI/CD pipeline for LangChain applications: mocked unit tests (free, fast), gated integration tests with real LLMs (costs money, slow), RAG pipeline validation, and LangSmith trace integration.
Sets up TypeScript LangChain dev workflow: project structure, Vitest config, mocked LLM unit tests with FakeListChatModel, integration tests.
Provides unit test mocks, integration tests with Testcontainers, and RAG validation patterns for LangChain4j Java AI applications. Use for testing AI services, retrieval chains, and LLM workflows.
Configures Langfuse in GitHub Actions for AI quality tests, prompt regression, tracing, and LLM monitoring in CI/CD pipelines.
Share bugs, ideas, or general feedback.
CI/CD pipeline for LangChain applications: mocked unit tests (free, fast), gated integration tests with real LLMs (costs money, slow), RAG pipeline validation, and LangSmith trace integration.
# .github/workflows/langchain-tests.yml
name: LangChain Tests
on:
pull_request:
paths: ["src/**", "tests/**", "package.json"]
jobs:
unit-tests:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
- uses: actions/setup-node@v4
with: { node-version: "20" }
- run: npm ci
- name: Unit tests (no API calls)
run: npx vitest run tests/unit/ --reporter=verbose
integration-tests:
runs-on: ubuntu-latest
if: github.event.pull_request.draft == false
needs: unit-tests
steps:
- uses: actions/checkout@v4
- uses: actions/setup-node@v4
with: { node-version: "20" }
- run: npm ci
- name: Integration tests (real LLM calls)
env:
OPENAI_API_KEY: ${{ secrets.OPENAI_API_KEY }}
LANGSMITH_TRACING: "true"
LANGSMITH_API_KEY: ${{ secrets.LANGSMITH_API_KEY }}
LANGSMITH_PROJECT: "ci-${{ github.run_id }}"
run: npx vitest run tests/integration/ --reporter=verbose
typecheck:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
- uses: actions/setup-node@v4
with: { node-version: "20" }
- run: npm ci
- run: npx tsc --noEmit
// tests/unit/chains.test.ts
import { describe, it, expect } from "vitest";
import { FakeListChatModel } from "@langchain/core/utils/testing";
import { ChatPromptTemplate } from "@langchain/core/prompts";
import { StringOutputParser } from "@langchain/core/output_parsers";
describe("Summarize Chain", () => {
const fakeLLM = new FakeListChatModel({
responses: ["Summary: LangChain enables LLM app development."],
});
it("produces output from prompt -> model -> parser", async () => {
const chain = ChatPromptTemplate.fromTemplate("Summarize: {text}")
.pipe(fakeLLM)
.pipe(new StringOutputParser());
const result = await chain.invoke({ text: "Long document..." });
expect(result).toContain("LangChain");
});
it("passes correct variables to prompt", () => {
const prompt = ChatPromptTemplate.fromTemplate("Translate {text} to {lang}");
expect(prompt.inputVariables).toContain("text");
expect(prompt.inputVariables).toContain("lang");
});
});
// tests/unit/tools.test.ts
import { describe, it, expect } from "vitest";
import { calculator, searchTool } from "../../src/tools";
describe("Calculator Tool", () => {
it("evaluates valid expressions", async () => {
expect(await calculator.invoke({ expression: "10 * 5" })).toBe("50");
});
it("returns error for invalid input", async () => {
const result = await calculator.invoke({ expression: "abc" });
expect(result).toContain("Error");
});
it("has correct metadata", () => {
expect(calculator.name).toBe("calculator");
expect(calculator.description).toBeTruthy();
});
});
// tests/integration/rag.test.ts
import { describe, it, expect } from "vitest";
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai";
import { MemoryVectorStore } from "langchain/vectorstores/memory";
import { ChatPromptTemplate } from "@langchain/core/prompts";
import { StringOutputParser } from "@langchain/core/output_parsers";
import { RunnableSequence, RunnablePassthrough } from "@langchain/core/runnables";
describe.skipIf(!process.env.OPENAI_API_KEY)("RAG Pipeline", () => {
it("retrieves relevant documents and answers correctly", async () => {
const embeddings = new OpenAIEmbeddings({ model: "text-embedding-3-small" });
const store = await MemoryVectorStore.fromTexts(
[
"LangChain was created by Harrison Chase in 2022.",
"LCEL stands for LangChain Expression Language.",
"Pinecone is a vector database for AI applications.",
],
[{}, {}, {}],
embeddings
);
const retriever = store.asRetriever({ k: 2 });
const model = new ChatOpenAI({ model: "gpt-4o-mini", temperature: 0 });
const prompt = ChatPromptTemplate.fromTemplate(
"Context: {context}\n\nQuestion: {question}\nAnswer:"
);
const chain = RunnableSequence.from([
{
context: retriever.pipe((docs) => docs.map((d) => d.pageContent).join("\n")),
question: new RunnablePassthrough(),
},
prompt,
model,
new StringOutputParser(),
]);
const answer = await chain.invoke("Who created LangChain?");
expect(answer.toLowerCase()).toContain("harrison");
});
it("handles questions outside context gracefully", async () => {
// Test that RAG doesn't hallucinate
const embeddings = new OpenAIEmbeddings({ model: "text-embedding-3-small" });
const store = await MemoryVectorStore.fromTexts(
["TypeScript is maintained by Microsoft."],
[{}],
embeddings
);
const retriever = store.asRetriever({ k: 1 });
const model = new ChatOpenAI({ model: "gpt-4o-mini", temperature: 0 });
const prompt = ChatPromptTemplate.fromTemplate(
"Based ONLY on this context, answer the question. Say 'I don't know' if not found.\n\nContext: {context}\n\nQuestion: {question}"
);
const chain = RunnableSequence.from([
{
context: retriever.pipe((docs) => docs.map((d) => d.pageContent).join("\n")),
question: new RunnablePassthrough(),
},
prompt,
model,
new StringOutputParser(),
]);
const answer = await chain.invoke("What is the capital of France?");
expect(answer.toLowerCase()).toMatch(/don.t know|not (in|found)|no information/);
});
});
# Gate integration tests behind PR labels or manual trigger
integration-tests:
if: |
github.event.pull_request.draft == false &&
contains(github.event.pull_request.labels.*.name, 'test:integration')
| Issue | Cause | Fix |
|---|---|---|
| Unit tests call real API | Didn't use FakeListChatModel | Replace ChatOpenAI with fake in tests |
| Integration test missing key | Secret not configured | Add OPENAI_API_KEY to repo secrets |
| Flaky RAG test | Embedding variability | Use deterministic data, set temperature: 0 |
| CI timeout | Model latency | Set timeout: 15000 on test, use gpt-4o-mini |
For deployment, see langchain-deploy-integration.