npx claudepluginhub dotnet/skills --plugin dotnet-aiThis skill uses the workspace's default tool permissions.
| Input | Required | Description |
Creates isolated Git worktrees for feature branches with prioritized directory selection, gitignore safety checks, auto project setup for Node/Python/Rust/Go, and baseline verification.
Executes implementation plans in current session by dispatching fresh subagents per independent task, with two-stage reviews: spec compliance then code quality.
Dispatches parallel agents to independently tackle 2+ tasks like separate test failures or subsystems without shared state or dependencies.
| Input | Required | Description |
|---|---|---|
| Task description | Yes | What the AI/ML feature should accomplish (e.g., "classify support tickets", "summarize documents") |
| Data description | Yes | Type and shape of input data (structured/tabular, unstructured text, images, mixed) |
| Deployment constraints | No | Cloud vs. local, latency SLO, cost budget, offline requirements |
| Existing project context | No | Current .csproj, existing packages, target framework |
Evaluate the developer's task against this decision tree and select the appropriate technology. State which branch applies and why.
| Task type | Technology | Rationale |
|---|---|---|
| Structured/tabular data: classification, regression, clustering, anomaly detection, recommendation | ML.NET (Microsoft.ML) | Reproducible (given a fixed seed and dataset), no cloud dependency, purpose-built models for these tasks |
| Natural language understanding, generation, summarization, reasoning over unstructured text (single prompt → response, no tool calling) | LLM via Microsoft.Extensions.AI (IChatClient) | Requires language model capabilities beyond pattern matching; no orchestration needed |
| Agentic workflows: tool/function calling, multi-step reasoning, agent loops, multi-agent collaboration | Microsoft Agent Framework (Microsoft.Agents.AI) built on top of Microsoft.Extensions.AI | Requires orchestration, tool dispatch, iteration control, and guardrails that IChatClient alone does not provide |
| Building GitHub Copilot extensions, custom agents, or developer workflow tools | GitHub Copilot SDK (GitHub.Copilot.SDK) | Integrates with the Copilot agent runtime for IDE and CLI extensibility |
| Running a pre-trained or fine-tuned custom model in production | ONNX Runtime (Microsoft.ML.OnnxRuntime) | Hardware-accelerated inference, model-format agnostic |
| Local/offline LLM inference with no cloud dependency | OllamaSharp with local AI models supported by Ollama | Privacy-sensitive, air-gapped, or cost-constrained scenarios |
| Semantic search, RAG, or embedding storage | Microsoft.Extensions.VectorData.Abstractions + a vector database provider (e.g., Azure AI Search, Milvus, MongoDB, pgvector, Pinecone, Qdrant, Redis, SQL) | Provider-agnostic abstractions for vector similarity search; pair with a database-specific connector package (many are moving to community toolkits) |
| Ingesting, chunking, and loading documents into a vector store | Microsoft.Extensions.AI.DataIngestion (preview) + Microsoft.Extensions.VectorData.Abstractions (MEVD) | Handles document parsing, text chunking, embedding generation, and upserting into a vector database; pairs with Microsoft.Extensions.VectorData.Abstractions |
| Both structured ML predictions AND natural language reasoning | Hybrid: ML.NET for predictions + LLM for reasoning layer | Keep loosely coupled; ML.NET handles reproducible scoring, LLM adds explanation |
Critical rule: Do NOT use an LLM for tasks that ML.NET handles well (classification on tabular data, regression, clustering). LLMs are slower, more expensive, and non-deterministic for these tasks.
After identifying the task type, select the right library layer. These libraries form a stack — each builds on the one below it. Using the wrong layer is a major source of non-deterministic agent behavior.
| Layer | Library | NuGet package | Use when |
|---|---|---|---|
| Abstraction | Microsoft.Extensions.AI (MEAI) | Microsoft.Extensions.AI | You need a provider-agnostic interface for chat, embeddings, or tool calling. This is the foundation — always include it. Use IChatClient directly only for simple prompt-in/response-out scenarios with no tool calling or agentic loops. If the task involves tools, agents, or multi-step reasoning, you must add the Orchestration layer above. |
| Provider SDK | OpenAI, Azure.AI.OpenAI, Azure.AI.Inference, OllamaSharp | OpenAI, Azure.AI.OpenAI, Azure.AI.Inference, OllamaSharp | You need a concrete LLM provider implementation. These wire into MEAI via AddChatClient. Use OpenAI for direct OpenAI access, Azure.AI.OpenAI for Azure OpenAI, Azure.AI.Inference for Azure AI Foundry / GitHub Models, or OllamaSharp for local Ollama. Use directly only if you need provider-specific features not exposed through MEAI. |
| Orchestration | Microsoft Agent Framework | Microsoft.Agents.AI (prerelease) | The task involves tool/function calling, agentic loops, multi-step reasoning, multi-agent coordination, durable context, or graph-based workflows. This is required whenever the scenario involves agents or tools — do not hand-roll tool dispatch loops with IChatClient. Builds on top of MEAI. Note: This package is currently prerelease — use dotnet add package Microsoft.Agents.AI --prerelease to install it. |
| Copilot integration | GitHub Copilot SDK | GitHub.Copilot.SDK | You are building extensions or tools that integrate with the GitHub Copilot runtime — custom agents, IDE extensions, or developer workflow automation that leverages the Copilot agent platform. |
Start with MEAI. Every AI integration begins with Microsoft.Extensions.AI for the IChatClient / IEmbeddingGenerator abstractions. This ensures provider-swappability and testability.
Add a provider SDK (OpenAI, Azure.AI.OpenAI) as the concrete implementation behind MEAI. Do not call the provider SDK directly in business logic — always go through the MEAI abstraction.
Use Agent Framework (Microsoft.Agents.AI) for any task that involves tools or agents. If the task is a single prompt → response with no tool calling, MEAI is sufficient. You MUST use Microsoft.Agents.AI when any of these apply:
Do not implement these patterns by hand with IChatClient — the Agent Framework provides iteration limits, observability, and tool dispatch that are error-prone to reimplement.
Add Copilot SDK only when building Copilot extensions. Use GitHub.Copilot.SDK when the goal is to build a custom agent or tool that runs inside the GitHub Copilot platform (CLI, IDE, or Copilot Chat). This is not a general-purpose LLM orchestration library — it is specifically for Copilot extensibility.
Never skip layers. Do not use Agent Framework without MEAI underneath. Do not call HttpClient to OpenAI alongside MEAI in the same workflow. Each layer depends on the one below it.
Install only the packages needed for the selected technology branch. Do not mix competing abstractions.
<PackageReference Include="Microsoft.ML" Version="4.*" />
<PackageReference Include="Microsoft.ML.AutoML" Version="0.*" />
<!-- Only if custom numerical work is needed: -->
PackageReference Include="System.Numerics.Tensors" Version="10.*"
<PackageReference Include="MathNet.Numerics" Version="5.*" />
<!-- Only for data exploration: -->
<PackageReference Include="Microsoft.Data.Analysis" Version="0.*" />
Do NOT use Accord.NET — it is archived and unmaintained.
<!-- Always start with the abstraction layer -->
<PackageReference Include="Microsoft.Extensions.AI" Version="9.*" />
<!-- Orchestration (agents, workflows, tools, memory) — prerelease; use dotnet add package Microsoft.Agents.AI --prerelease -->
<PackageReference Include="Microsoft.Agents.AI" Version="1.*-*" />
<!-- Cloud LLM provider (pick one) -->
<PackageReference Include="Azure.AI.OpenAI" Version="2.*" />
<!-- OR -->
<PackageReference Include="OpenAI" Version="2.*" />
<!-- Client-side token counting for cost management -->
<PackageReference Include="Microsoft.ML.Tokenizers" Version="2.*"
<!-- Local LLM inference -->
<PackageReference Include="OllamaSharp" Version="5.*" />
<!-- Custom model inference -->
<PackageReference Include="Microsoft.ML.OnnxRuntime" Version="1.*" />
<!-- Vector store abstraction -->
<PackageReference Include="Microsoft.Extensions.VectorData.Abstractions" Version="9.*" />
<!-- Document ingestion, chunking, and vector store loading (preview) -->
<PackageReference Include="Microsoft.Extensions.AI.DataIngestion" Version="9.*-*" />
<!-- Copilot platform extensibility -->
<PackageReference Include="GitHub.Copilot.SDK" Version="1.*" />
Stack coherence rule: Never mix raw SDK calls (
HttpClientto OpenAI) withMicrosoft.Extensions.AI, Microsoft Agent Framework, or Copilot SDK in the same workflow. Pick one abstraction layer per workflow boundary and commit to it. See Step 1b for the layering rules.
All AI/ML services must be registered via DI. Never instantiate clients directly in business logic.
// Configuration via IOptions<T>
services.Configure<AiOptions>(configuration.GetSection("AI"));
// Register the AI client through the abstraction
services.AddChatClient(builder => builder
.UseOpenAIChatClient("gpt-4o-mini-2024-07-18"));
Apply the guardrails for the selected technology branch. Every generated implementation must follow these rules.
Reproducibility: Always set a random seed in the ML context:
var mlContext = new MLContext(seed: 42);
Data splitting: Always split into train/test (and optionally validation). Never evaluate on training data:
var split = mlContext.Data.TrainTestSplit(data, testFraction: 0.2);
Metrics logging: Always compute and log evaluation metrics appropriate to the task:
var metrics = mlContext.BinaryClassification.Evaluate(predictions);
logger.LogInformation("AUC: {Auc:F4}, F1: {F1:F4}", metrics.AreaUnderRocCurve, metrics.F1Score);
AutoML first: Prefer mlContext.Auto() for initial model selection, then refine manually.
PredictionEngine pooling: In ASP.NET Core, always use the pooled prediction engine — never a singleton:
services.AddPredictionEnginePool<ModelInput, ModelOutput>()
.FromFile(modelPath);
Temperature: Always set explicitly. Use 0 for factual/deterministic tasks:
var options = new ChatOptions
{
Temperature = 0f,
MaxOutputTokens = 1024,
};
Structured output: Always parse LLM output into strongly-typed objects with fallback handling:
var result = await chatClient.GetResponseAsync<MySchema>(prompt, options, cancellationToken);
Retry logic: Always implement retry with exponential backoff:
services.AddChatClient(builder => builder
.UseOpenAIChatClient(modelId)
.Use(new RetryingChatClient(maxRetries: 3)));
Cost control: Always estimate and log token usage. Use Microsoft.ML.Tokenizers to count tokens client-side before sending requests so you can enforce budgets proactively. Choose the smallest model tier that meets quality requirements (e.g., gpt-4o-mini before gpt-4o).
Secret management: Never hardcode API keys. Use Azure Key Vault, user-secrets, or environment variables:
var apiKey = configuration["AI:ApiKey"]
?? throw new InvalidOperationException("AI:ApiKey not configured");
Model version pinning: Specify exact model versions to reduce behavioral drift:
// Pin to a specific dated version, not just "gpt-4o"
var modelId = "gpt-4o-2024-08-06";
Use Microsoft.Agents.AI for all agentic workflows. Do not implement tool dispatch loops or multi-step agent reasoning by hand with IChatClient. The Agent Framework provides ChatClientAgent (or AgentWorker) which handles the tool call → result → re-prompt cycle with built-in guardrails. All rules below assume you are using Microsoft.Agents.AI.
Iteration limits: Always cap agentic loops to prevent runaway execution:
var settings = new AgentInvokeOptions
{
MaximumIterations = 10,
};
Cost ceiling: Implement a token budget per execution and terminate when reached. Use Microsoft.ML.Tokenizers to count prompt and completion tokens locally and compare against the budget before each iteration.
Observability: Log non-sensitive metadata for every agent step. Never log raw message.Content — it may contain user prompts, tool outputs, secrets, or PII that persist in plaintext in central logging systems:
await foreach (var message in agent.InvokeStreamingAsync(history, settings))
{
logger.LogDebug("Agent step: Role={Role}, ContentLength={Length}",
message.Role, message.Content?.Length ?? 0);
}
Tool schemas: Define explicit tool/function schemas with descriptions. Never rely on implicit tool discovery.
Simplicity preference: Prefer single-agent with tools over multi-agent unless the task genuinely requires agent collaboration.
Embedding caching: Never re-embed the same content on every query. Cache embeddings in the vector store.
Chunking strategy: Use semantic chunking (split on paragraph/section boundaries) over fixed-size chunking. Ensure chunks have enough context to be useful on their own.
Relevance thresholds: Do not inject low-relevance chunks into context. Set a minimum similarity score:
var results = await vectorStore.SearchAsync(query, new VectorSearchOptions
{
Top = 5,
MinimumScore = 0.75f,
});
Source attribution: Track which chunks contributed to the final response. Include source references in the output.
Batch embeddings: Batch embedding API calls where possible to reduce latency and cost.
When the solution involves LLM calls or agentic workflows, explicitly address non-determinism:
Acknowledge it: Inform the developer that LLM outputs are non-deterministic even at temperature 0 (due to batching, quantization, and model updates).
Validate outputs: Implement schema validation and content assertion checks on every LLM response.
Graceful degradation: Design a fallback path for when the LLM returns unexpected, malformed, or empty output:
var response = await chatClient.GetResponseAsync<ClassificationResult>(prompt, options);
if (response is null || !response.IsValid())
{
logger.LogWarning("LLM returned invalid response, falling back to rule-based classifier");
return ruleBasedClassifier.Classify(input);
}
Evaluation harness: For any prompt that will be iterated on, recommend creating a golden dataset and evaluation scaffold to measure prompt quality over time.
Model version pinning: Pin to specific dated model versions (e.g., gpt-4o-2024-08-06) to reduce drift between deployments.
Connection pooling: Use IHttpClientFactory and DI-managed clients for all external services.
Response caching: Cache repeated or similar queries. Consider semantic caching for LLM responses where appropriate.
Streaming: Use IAsyncEnumerable for LLM responses in user-facing scenarios to reduce time-to-first-token:
await foreach (var update in chatClient.GetStreamingResponseAsync(prompt, options))
{
yield return update.Text;
}
Health checks: Implement health checks for external AI service dependencies:
services.AddHealthChecks()
.AddCheck<OpenAIHealthCheck>("openai");
ML.NET prediction pooling: In web applications, always use PredictionEnginePool<TIn, TOut>, never a single PredictionEngine instance (it is not thread-safe).
Build the project and verify no warnings:
dotnet build -c Release -warnaserror
Run tests, including integration tests that validate AI/ML behavior:
dotnet test -c Release
For ML.NET pipelines, verify that evaluation metrics meet the project's quality bar and that the model can be serialized and loaded correctly.
For LLM integrations, verify that structured output parsing handles both valid and malformed responses.
For RAG pipelines, verify that retrieval returns relevant results and that irrelevant chunks are filtered out.
IOptions<T> pattern — no hardcoded valuesdotnet build -c Release -warnaserror completes cleanlyWhen reviewing or generating code, flag and redirect the developer if any of these patterns are detected:
| Anti-pattern | Redirect |
|---|---|
| Using an LLM for classification on structured/tabular data | Use ML.NET instead — it is faster, cheaper, and deterministic |
| Calling LLM APIs without retry or timeout logic | Add RetryingChatClient or Polly-based retry with exponential backoff |
Storing API keys in appsettings.json committed to source control | Use user-secrets (dev), environment variables, or Azure Key Vault (prod) |
| Using Accord.NET for new projects | Migrate to ML.NET — Accord.NET is archived and unmaintained |
| Building custom neural networks in .NET from scratch | Use a pre-trained model via ONNX Runtime or call an LLM API |
| RAG without chunking strategy or relevance filtering | Implement semantic chunking and set a minimum similarity score threshold |
| Agentic loops without iteration limits or cost ceilings | Add MaximumIterations and a token budget ceiling |
Using MEAI IChatClient with raw HttpClient calls to the same provider | Pick one abstraction layer and commit to it |
Implementing tool calling or agentic loops manually with IChatClient instead of using Microsoft.Agents.AI | Use Microsoft.Agents.AI — it provides iteration limits (MaximumIterations), built-in tool dispatch, observability hooks, and cost controls. Hand-rolled loops lack these guardrails. |
| Using Agent Framework for a single prompt→response call | Use MEAI IChatClient directly — Agent Framework is for multi-step orchestration |
| Using Copilot SDK for general-purpose LLM apps | Copilot SDK is for Copilot platform extensions only — use MEAI + Agent Framework for standalone apps |
| Calling OpenAI SDK directly in business logic instead of through MEAI | Register the provider via AddChatClient and depend on IChatClient in business code |
Using PredictionEngine as a singleton in ASP.NET Core | Use PredictionEnginePool<TIn, TOut> — PredictionEngine is not thread-safe |
Using Func<ReadOnlySpan<T>> for delegates with ref struct parameters | Define a custom delegate type — ref structs cannot be generic type arguments |
Using Microsoft.SemanticKernel for new projects | Use Microsoft.Extensions.AI + Microsoft.Agents.AI — Semantic Kernel is superseded by these newer abstractions for LLM orchestration and tool calling |
| Pitfall | Solution |
|---|---|
| Over-engineering with LLMs | Start with the simplest approach (rules, ML.NET) and add LLM capability only when simpler methods fall short |
| Evaluating ML models on training data | Always use TrainTestSplit and report metrics on the held-out test set |
| LLM output drift between deployments | Pin to specific dated model versions (e.g., gpt-4o-2024-08-06) |
| Token cost surprises | Set MaxOutputTokens, use Microsoft.ML.Tokenizers for accurate client-side token counting, log token counts per request, and alert on budget thresholds |
| Non-reproducible ML training | Set MLContext(seed: N) and version your training data alongside the code |
| RAG returning irrelevant context | Set a minimum similarity score and limit the number of injected chunks |
| Cold start latency on ML.NET models | Pre-warm the PredictionEnginePool during application startup |
| Microsoft Agent Framework + raw OpenAI SDK in same class | Choose one orchestration layer per workflow boundary |