Guides implementation of AI system architectures — technology selection, pipeline implementation, model serving setup, monitoring deployment, and CI/CD automation. This skill should be used when the user asks to "implement AI architecture", "build ML pipeline", "set up model serving", "deploy AI system", "implement MLOps", "configure drift monitoring", "set up feature store", or mentions AI implementation plan, ML infrastructure setup, model deployment guide, RAG implementation, or agent framework setup.
From maonpx claudepluginhub javimontano/mao-discovery-frameworkThis skill is limited to using the following tools:
references/implementation-blueprints.mdreferences/implementation-playbook.mdreferences/technology-selection.mdIntegrates Apple's FoundationModels for on-device LLM in iOS 26+ apps: text generation, @Generable structured output, tool calling, snapshot streaming.
Provides React and Next.js patterns for component composition, compound components, state management, data fetching, performance optimization, forms, routing, and accessible UIs.
Reviews Flutter/Dart code with library-agnostic checklist for widget best practices, state management patterns, Dart idioms, performance, accessibility, security, and clean architecture.
Guiar la implementación de arquitecturas AI desde el diseño hasta producción — selección de tecnología, implementación de pipelines, configuración de serving, despliegue de monitoreo, y automatización CI/CD. Produce blueprints de implementación, guías de selección de tecnología, y un playbook fase-a-fase que transforma decisiones arquitectónicas en infraestructura operativa.
Phased delivery, not big bang. Implementar en fases con valor entregable por fase. Fase 0 (Foundation) → Fase 1 (Data Pipeline) → Fase 2 (Model Development) → Fase 3 (Serving) → Fase 4 (CI/CD) → Fase 5 (Monitoring). Cada fase produce capacidad usable.
Start simple, evolve with evidence. No sobre-ingenierar para escala hipotética. Empezar con la implementación más simple que resuelva el problema actual. Feature Store no es necesario para un modelo; multi-model tiering no es necesario para un tier.
Tests and monitoring from Phase 0, not Phase 5. La infraestructura de testing y monitoreo se establece en la primera fase, no se agrega después de incidentes. Cada componente implementado incluye sus tests y sus métricas desde el día uno.
Parámetros:
MODO: [greenfield | brownfield | remediation | migration]
FORMATO: [ejecutivo | técnico | híbrido]
VARIANTE: [pipeline | serving | genai | mlops | full]
STACK: [python | java | typescript | polyglot]
INFRA: [kubernetes | serverless | containers | hybrid]
Detección automática:
- Si existe codebase → MODO=brownfield
- Si existe audit report → MODO=remediation
- Si el input menciona "migrar" → MODO=migration
- Si existe LangChain/LlamaIndex → VARIANTE incluye genai
- Default: MODO=greenfield, VARIANTE=full, STACK=python, INFRA=containers
Selecciona las tecnologías para cada componente del sistema AI con justificación basada en constraints.
Load references:
Read ${CLAUDE_SKILL_DIR}/references/technology-selection.md
Decision Framework:
Selection matrix por componente:
| Componente | Tecnología | Alternativa | Justificación |
|---|---|---|---|
| ML Framework | [selección] | [alternativa] | [por qué] |
| Pipeline Orchestration | [selección] | [alternativa] | [por qué] |
| Experiment Tracking | [selección] | [alternativa] | [por qué] |
| Model Registry | [selección] | [alternativa] | [por qué] |
| Model Serving | [selección] | [alternativa] | [por qué] |
| Vector DB (si GenAI) | [selección] | [alternativa] | [por qué] |
| LLM Framework (si GenAI) | [selección] | [alternativa] | [por qué] |
| Data Quality | [selección] | [alternativa] | [por qué] |
| Monitoring | [selección] | [alternativa] | [por qué] |
| CI/CD | [selección] | [alternativa] | [por qué] |
Entregable: Stack decision document con ADRs por cada selección no-obvia.
Implementa el pipeline de datos desde ingestion hasta feature serving.
Load references:
Read ${CLAUDE_SKILL_DIR}/references/implementation-blueprints.md
Componentes a implementar:
Blueprint selection:
Entregable: Implemented pipeline with tests, monitoring, and documentation.
Implementa el ciclo de desarrollo de modelos con tracking, registro, y evaluación.
Componentes a implementar:
Blueprint selection:
Entregable: Training pipeline, experiment tracking, model registry operational.
Implementa model serving, API layer, caching, y fallback mechanisms.
Componentes a implementar:
GenAI-specific implementation:
Entregable: Serving infrastructure operational, API documented, load tested.
Implementa Blue & Gold deployment con validation gates automatizados.
Componentes a implementar:
Gate configuration:
Entregable: CI/CD pipeline operational, gates configured, rollback tested.
Implementa el stack de observabilidad completo para el sistema AI.
Componentes a implementar:
Dashboard hierarchy:
Entregable: Full observability stack, dashboards, alerts, runbooks.
| Decision | Enables | Constrains | When to Use |
|---|---|---|---|
| Managed MLOps platform | Fast setup, integrated tools | Vendor lock-in, less customization | Teams without ML infra expertise |
| Open-source stack | Full control, no lock-in | Integration effort, operations burden | Teams with strong infra skills |
| Monorepo | Unified CI/CD, shared code | Build complexity, repo size | Small-medium teams, shared components |
| Multi-repo | Team autonomy, independent releases | Integration testing harder, code duplication | Large teams, independent services |
| Kubernetes-native | Scaling, orchestration, ecosystem | Complexity, K8s expertise required | Multi-model, high-scale systems |
| Serverless | Zero-ops, pay-per-use | Cold starts, limited customization | Event-driven, low-to-medium traffic |
Equipo sin experiencia ML: Fase 0 extendida con capacitación. Empezar con managed services (SageMaker, Vertex). Reducir complejidad de Feature Store y multi-model tiering hasta que el equipo madure.
Migración desde notebooks: Priorizar extracción de feature engineering y training logic a módulos Python testeables. Notebooks quedan solo para exploración. Fase 2 se convierte en la fase más larga.
Remediación post-auditoría: Ordenar implementación por priority score del audit report, no por la secuencia estándar de fases. Puede requerir empezar por Fase 5 (monitoring) si el hallazgo crítico es "no hay observabilidad".
Sistema GenAI puro (sin ML tradicional): Fases 2 y 3 se fusionan en "RAG/Agent Implementation". Feature Store no aplica. Focus en guardrails, vector DB, prompt management, cost controls.
Restricciones de presupuesto extremas: Open-source everything. MLflow (free), Feast (free), Evidently (free), GitHub Actions (free tier). Docker Compose para desarrollo, single-instance para producción inicial.
| Skill | Relación |
|---|---|
ai-software-architecture | Proporciona diseño de arquitectura a implementar |
ai-pipeline-architecture | Proporciona diseño de pipelines a implementar |
ai-design-patterns | Proporciona patrones seleccionados a implementar |
ai-testing-strategy | Proporciona estrategia de testing a implementar en CI/CD |
genai-architecture | Proporciona diseño GenAI a implementar |
ai-architecture-audit | Proporciona roadmap de remediación a ejecutar |
ai-conops | Proporciona modos operacionales y métricas de éxito |
aws-architecture-implementation | Implementación AWS-específica (complementaria) |
aws-architecture-design | Diseño AWS que se implementa con el skill AWS |
infrastructure-architecture | Proporciona diseño de infraestructura base |
if FORMATO == "ejecutivo":
Stack summary + implementation timeline + resource needs + milestones
Audiencia: Project managers, sponsors
if FORMATO == "técnico":
Full 6-section implementation guide + blueprints + configurations
Audiencia: ML engineers, DevOps, platform engineers
if FORMATO == "híbrido":
Executive timeline + technical deep-dive completo
Audiencia: Tech leads planning implementation sprints
## {System Name} — AI Architecture Implementation Guide
### Stack Decision
[S1: Technology selection matrix with ADRs]
### Phase 0: Foundation
[Repository structure, CI skeleton, development environment]
### Phase 1: Data Pipeline
[S2: Pipeline implementation, quality gates, feature store]
### Phase 2: Model Development
[S3: Training pipeline, experiment tracking, model registry]
### Phase 3: Serving & Inference
[S4: Model serving, API, caching, guardrails, RAG/agents]
### Phase 4: CI/CD & Deployment
[S5: Blue & Gold pipeline, gates, canary, rollback]
### Phase 5: Monitoring & Operations
[S6: Observability stack, dashboards, alerts, runbooks]
### Implementation Timeline
[Gantt chart with milestones and dependencies]