AI pipeline architecture design — development pipelines, production pipelines, data stores, model registry, CI/CD for AI, and non-functional requirements. This skill should be used when the user asks to "design AI pipelines", "architect ML pipelines", "select data stores for AI", "design model registry", "implement CI/CD for ML", "define AI pipeline requirements", or mentions MLOps, training pipeline, inference pipeline, feature pipeline, Blue and Gold deployment, or pipeline patterns. [EXPLICIT]
From jm-adknpx claudepluginhub javimontano/jm-adk-alfaThis skill is limited to using the following tools:
agents/guardian.mdagents/lead.mdagents/specialist.mdagents/support.mdevals/evals.jsonknowledge/body-of-knowledge.mdknowledge/knowledge-graph.mdprompts/meta.mdprompts/primary.mdprompts/variations/deep.mdprompts/variations/quick.mdreferences/data-stores.mdreferences/pipeline-patterns.mdreferences/requirements-tables.mdtemplates/output.docx.mdtemplates/output.htmlAI pipeline architecture defines how data flows through AI systems — from raw ingestion through model training and serving to production monitoring. This skill produces comprehensive pipeline architecture documentation covering development pipelines (experimentation to model artifact), production pipelines (data ingestion to prediction delivery), data store selection, model registry design, CI/CD strategy, and measurable requirements. [EXPLICIT]
El pipeline es la arquitectura. El modelo es solo un componente dentro de él. La mayoría del esfuerzo en sistemas de IA de producción está en la infraestructura de datos, no en el algoritmo. Un pipeline mal diseñado convierte un buen modelo en un sistema frágil. Un pipeline bien diseñado permite que modelos mediocres evolucionen.
The user provides a system or project name as $ARGUMENTS. Parse $1 as the system/project name used throughout all output artifacts. [EXPLICIT]
Parameters:
{MODO}: piloto-auto (default) | desatendido | supervisado | paso-a-paso{FORMATO}: markdown (default) | html | dual{ALCANCE}: ejecutiva (~40% — S1 dev pipeline + S2 prod pipeline + S5 CI/CD) | tecnica (full 6 sections, default)Before generating architecture, detect the codebase context:
Detección automática de contexto:
Escanear el codebase por frameworks ML (PyTorch, TensorFlow, scikit-learn),
orquestadores (Airflow, Dagster, Prefect, Kubeflow), y serving frameworks
(TensorFlow Serving, TorchServe, Triton, vLLM) para adaptar recomendaciones. [EXPLICIT]
Load references:
Read ${CLAUDE_SKILL_DIR}/references/pipeline-patterns.md
Read ${CLAUDE_SKILL_DIR}/references/data-stores.md
Read ${CLAUDE_SKILL_DIR}/references/requirements-tables.md
Maps the experimentation-to-artifact pipeline where models are built, trained, and validated. [EXPLICIT]
Stages:
Key decisions:
Maps the data-to-prediction pipeline that serves AI capabilities in production. [EXPLICIT]
Stages:
Key decisions:
Selects appropriate storage technologies for each pipeline component. [EXPLICIT]
Store types and AI use cases:
Selection criteria: Query complexity, latency requirements, scale, consistency model, cost, AI-native capability.
Multi-store pattern: Most production AI systems combine 3-4 store types with synchronization and lineage tracking across boundaries.
Designs the bridge between development and production pipelines. [EXPLICIT]
Registry capabilities:
Key decisions:
Designs the deployment strategy connecting development artifacts to production serving. [EXPLICIT]
Blue Pipeline (Production): Currently serving, fully validated, rollback target. Gold Pipeline (Staging): New version under validation, receives shadow/canary traffic.
Validation gates:
Promotion flow: Gold passes all gates → canary traffic → gradual shift → full promotion → Gold becomes Blue → previous Blue archived.
Key decisions:
Defines measurable requirements across four categories with thresholds and objectives. [EXPLICIT]
Performance (AP-1 to AP-13): Data processing speed, model accuracy, fairness, explainability, robustness.
Non-Functional (NF-1 to NF-9): Availability (>99.9%), recovery time (<1 min), fault detection (<0.5 secs), drift detection (<1 hour), pipeline isolation.
Security (SEC-1 to SEC-6): PKI for external interfaces, audit logging, adversarial protection, data access controls, model extraction monitoring.
Compliance (CP-1 to CP-7): Authorized data access, transaction archival, encryption at rest/in use, audit trails, model governance workflows.
| Decision | Enables | Constrains | When to Use |
|---|---|---|---|
| Batch Pipeline | Simple, cost-effective, easy debugging | High latency, not real-time | Offline analytics, nightly retraining |
| Streaming Pipeline | Real-time predictions, low latency | Complex, exactly-once semantics hard | Real-time fraud, recommendations |
| Hybrid Pipeline | Best of both, flexible | Two systems to maintain, consistency | Most production AI systems |
| Feature Store | Consistency, reuse, drift monitoring | Infra overhead, governance cost | Multiple models sharing features |
| Blue & Gold CI/CD | Safe deployments, instant rollback | Doubled infrastructure during validation | All production AI systems |
| Single Model Registry | Central governance, clear lineage | Single point of failure, access bottleneck | Standard team size |
| Distributed Registry | Team autonomy, reduced bottleneck | Consistency challenges, governance complexity | Large multi-team orgs |
Notebook-to-Production Migration: Data scientists work in Jupyter notebooks; production requires orchestrated pipelines. Bridge with notebook-aware orchestrators (Papermill, Ploomber). Prioritize extracting feature engineering and model training into reusable pipeline stages. [EXPLICIT]
Multi-Team Pipeline Ownership: Different teams own different pipeline stages (data eng owns ingestion, ML eng owns training, platform owns serving). Clear data contracts between stages are essential. Feature store becomes the coordination point. [EXPLICIT]
Real-Time + Batch Hybrid: System needs both real-time predictions (online serving) and batch analytics (offline scoring). Lambda or Kappa architecture patterns. Feature store must support both online (low-latency) and offline (batch) serving. [EXPLICIT]
Regulated Pipeline (Finance, Healthcare): Every pipeline stage must produce audit-worthy artifacts. Data lineage tracking from source to prediction. Model governance gates require human approval before production promotion. [EXPLICIT]
piloto-auto (default).Before finalizing delivery, verify:
El agente que ejecuta este skill debe verificar cada item antes de entregar el output al usuario.
| Format | Default | Description |
|---|---|---|
markdown | Yes | Rich Markdown + Mermaid diagrams. Token-efficient. |
html | On demand | Branded HTML (Design System). Visual impact. |
dual | On demand | Both formats. |
Primary: A-02_AI_Pipeline_Architecture_Deep.html — Development pipeline diagram, production pipeline diagram, data store selection matrix, model registry design, Blue & Gold CI/CD flow, requirements framework tables.
Secondary: Pipeline stage contracts (.md), data store comparison matrix, model registry workflow diagram (Mermaid/PNG/SVG), requirements checklist.
Fuente: Avila, R.D. & Ahmad, I. (2025). Architecting AI Software Systems. Packt.
Searches, retrieves, and installs Agent Skills from prompts.chat registry using MCP tools like search_skills and get_skill. Activates for finding skills, browsing catalogs, or extending Claude.