AI pipeline architecture design — development pipelines, production pipelines, data stores, model registry, CI/CD for AI, and non-functional requirements. This skill should be used when the user asks to "design AI pipelines", "architect ML pipelines", "select data stores for AI", "design model registry", "implement CI/CD for ML", "define AI pipeline requirements", or mentions MLOps, training pipeline, inference pipeline, feature pipeline, Blue and Gold deployment, or pipeline patterns.
From maonpx claudepluginhub javimontano/mao-discovery-frameworkThis skill is limited to using the following tools:
references/data-stores.mdreferences/pipeline-patterns.mdreferences/requirements-tables.mdIntegrates Apple's FoundationModels for on-device LLM in iOS 26+ apps: text generation, @Generable structured output, tool calling, snapshot streaming.
Provides React and Next.js patterns for component composition, compound components, state management, data fetching, performance optimization, forms, routing, and accessible UIs.
Reviews Flutter/Dart code with library-agnostic checklist for widget best practices, state management patterns, Dart idioms, performance, accessibility, security, and clean architecture.
AI pipeline architecture defines how data flows through AI systems — from raw ingestion through model training and serving to production monitoring. This skill produces comprehensive pipeline architecture documentation covering development pipelines (experimentation to model artifact), production pipelines (data ingestion to prediction delivery), data store selection, model registry design, CI/CD strategy, and measurable requirements.
El pipeline es la arquitectura. El modelo es solo un componente dentro de el. La mayoria del esfuerzo en sistemas de IA de produccion esta en la infraestructura de datos, no en el algoritmo. Un pipeline mal disenado convierte un buen modelo en un sistema fragil. Un pipeline bien disenado permite que modelos mediocres evolucionen.
The user provides a system or project name as $ARGUMENTS. Parse $1 as the system/project name used throughout all output artifacts.
Parameters:
{MODO}: piloto-auto (default) | desatendido | supervisado | paso-a-paso{FORMATO}: markdown (default) | html | dual{VARIANTE}: ejecutiva (~40% — S1 dev pipeline + S2 prod pipeline + S5 CI/CD) | tecnica (full 6 sections, default)Before generating architecture, detect the codebase context:
!find . -name "*.py" -o -name "Dockerfile" -o -name "*.yaml" -o -name "*.yml" | head -30
Detect ML frameworks (PyTorch, TensorFlow, scikit-learn), orchestrators (Airflow, Dagster, Prefect, Kubeflow), and serving frameworks (TensorFlow Serving, TorchServe, Triton, vLLM).
If reference materials exist, load them:
Read ${CLAUDE_SKILL_DIR}/references/pipeline-patterns.md
Read ${CLAUDE_SKILL_DIR}/references/data-stores.md
Read ${CLAUDE_SKILL_DIR}/references/requirements-tables.md
Maps the experimentation-to-artifact pipeline where models are built, trained, and validated.
Stages:
Key decisions:
Maps the data-to-prediction pipeline that serves AI capabilities in production.
Stages:
Key decisions:
Selects appropriate storage technologies for each pipeline component.
Store types and AI use cases:
Selection criteria: Query complexity, latency requirements, scale, consistency model, cost, AI-native capability.
Multi-store pattern: Most production AI systems combine 3-4 store types with synchronization and lineage tracking across boundaries.
Designs the bridge between development and production pipelines.
Registry capabilities:
Key decisions:
Designs the deployment strategy connecting development artifacts to production serving.
Blue Pipeline (Production): Currently serving, fully validated, rollback target. Gold Pipeline (Staging): New version under validation, receives shadow/canary traffic.
Validation gates:
Promotion flow: Gold passes all gates → canary traffic → gradual shift → full promotion → Gold becomes Blue → previous Blue archived.
Key decisions:
Defines measurable requirements across four categories with thresholds and objectives.
Performance (AP-1 to AP-13): Data processing speed, model accuracy, fairness, explainability, robustness.
Non-Functional (NF-1 to NF-9): Availability (>99.9%), recovery time (<1 min), fault detection (<0.5 secs), drift detection (<1 hour), pipeline isolation.
Security (SEC-1 to SEC-6): PKI for external interfaces, audit logging, adversarial protection, data access controls, model extraction monitoring.
Compliance (CP-1 to CP-7): Authorized data access, transaction archival, encryption at rest/in use, audit trails, model governance workflows.
| Decision | Enables | Constrains | When to Use |
|---|---|---|---|
| Batch Pipeline | Simple, cost-effective, easy debugging | High latency, not real-time | Offline analytics, nightly retraining |
| Streaming Pipeline | Real-time predictions, low latency | Complex, exactly-once semantics hard | Real-time fraud, recommendations |
| Hybrid Pipeline | Best of both, flexible | Two systems to maintain, consistency | Most production AI systems |
| Feature Store | Consistency, reuse, drift monitoring | Infra overhead, governance cost | Multiple models sharing features |
| Blue & Gold CI/CD | Safe deployments, instant rollback | Doubled infrastructure during validation | All production AI systems |
| Single Model Registry | Central governance, clear lineage | Single point of failure, access bottleneck | Standard team size |
| Distributed Registry | Team autonomy, reduced bottleneck | Consistency challenges, governance complexity | Large multi-team orgs |
Notebook-to-Production Migration: Data scientists work in Jupyter notebooks; production requires orchestrated pipelines. Bridge with notebook-aware orchestrators (Papermill, Ploomber). Prioritize extracting feature engineering and model training into reusable pipeline stages.
Multi-Team Pipeline Ownership: Different teams own different pipeline stages (data eng owns ingestion, ML eng owns training, platform owns serving). Clear data contracts between stages are essential. Feature store becomes the coordination point.
Real-Time + Batch Hybrid: System needs both real-time predictions (online serving) and batch analytics (offline scoring). Lambda or Kappa architecture patterns. Feature store must support both online (low-latency) and offline (batch) serving.
Regulated Pipeline (Finance, Healthcare): Every pipeline stage must produce audit-worthy artifacts. Data lineage tracking from source to prediction. Model governance gates require human approval before production promotion.
Before finalizing delivery, verify:
| Format | Default | Description |
|---|---|---|
markdown | Yes | Rich Markdown + Mermaid diagrams. Token-efficient. |
html | On demand | Branded HTML (Design System). Visual impact. |
dual | On demand | Both formats. |
Primary: A-02_AI_Pipeline_Architecture_Deep.html — Development pipeline diagram, production pipeline diagram, data store selection matrix, model registry design, Blue & Gold CI/CD flow, requirements framework tables.
Secondary: Pipeline stage contracts (.md), data store comparison matrix, model registry workflow diagram (Mermaid/PNG/SVG), requirements checklist.
Fuente: Avila, R.D. & Ahmad, I. (2025). Architecting AI Software Systems. Packt.