Help us improve
Share bugs, ideas, or general feedback.
Chief AI Officer advisory for startups: model build-vs-buy decisions (API vs fine-tune vs in-house), AI risk classification under EU AI Act + US state patchwork, AI cost economics (API-to-self-hosted breakeven), and AI team org evolution. Use when deciding whether to call an API or fine-tune, classifying AI use cases for regulatory risk, calculating when self-hosting pays off, sequencing AI hires, or when user mentions CAIO, AI strategy, model selection, foundation model, fine-tuning, EU AI Act, NIST AI RMF, AI governance, model risk, or AI economics. Strategic only — does not duplicate engineering AI/ML skills.
npx claudepluginhub ciciliaeth/claude-skills --plugin chief-ai-officer-advisorHow this skill is triggered — by the user, by Claude, or both
Slash command
/chief-ai-officer-advisor:chief-ai-officer-advisorThe summary Claude sees in its skill listing — used to decide when to auto-load this skill
Strategic AI leadership for startup CAIOs and founders without one. **Four decisions, no AI hype:**
Provides PHI/PII compliance patterns for healthcare apps including data classification, row-level security access control, audit trails, encryption, and common leak vectors. Useful for patient data features, APIs, and code reviews.
Share bugs, ideas, or general feedback.
Strategic AI leadership for startup CAIOs and founders without one. Four decisions, no AI hype:
This skill does not cover tactical AI/ML engineering. For RAG implementation, agent design, prompt engineering, eval infrastructure, model deployment, or cost optimization, see engineering/rag-architect/, engineering/agent-designer/, engineering/prompt-governance/, engineering/self-eval/, engineering/llm-cost-optimizer/.
CAIO, chief AI officer, AI strategy, model selection, foundation model, fine-tuning, RLHF, DPO, LoRA, QLoRA, build vs buy, AI build-vs-buy, model risk tier, EU AI Act, AI Act Article 6, Article 9, Article 10, Annex III, prohibited AI, high-risk AI, NIST AI RMF, AI risk management framework, NYC Local Law 144, Colorado SB 21-169, Illinois HB 53, model card, eval set, eval harness, hallucination rate, jailbreak risk, prompt injection, AI red team, AI safety, alignment, model lifecycle, model registry, API-to-self-hosted breakeven, GPU economics, A100, H100, inference cost, fine-tuning cost, AI team, AI engineer, ML engineer, research scientist, MLOps, AI platform
# Decision A: API vs fine-tune vs build
python scripts/model_buildvsbuy_calculator.py # embedded customer-support sample
python scripts/model_buildvsbuy_calculator.py path/to/use_case.json
# Decision B: Risk classification under EU AI Act + US state laws
python scripts/ai_risk_classifier.py # embedded hiring-AI sample
python scripts/ai_risk_classifier.py path/to/use_case.json
# Decision C: API vs self-hosted economics
python scripts/ai_cost_economics.py # embedded 5M tokens/day sample
python scripts/ai_cost_economics.py path/to/workload.json
The decision is not "use AI or not" — it's API vs fine-tune vs in-house for each use case. Each path has a different TCO curve, latency profile, and capability ceiling.
Default path: API (frontier model)
Fine-tune a smaller model
Build from scratch / pre-train
Run model_buildvsbuy_calculator.py for a use-case-specific recommendation with 3-year TCO. See references/model_buildvsbuy_strategy.md for full decision tree.
The 2026 question every founder is facing: does this AI use case trigger high-risk regulatory obligations?
EU AI Act (in force 2026) tiers:
| Tier | Examples | Obligations |
|---|---|---|
| Prohibited | Social scoring, real-time biometric surveillance, manipulative AI | Cannot deploy in EU |
| High-risk | Employment screening, credit scoring, education access, critical infrastructure, law enforcement, biometric ID | Conformity assessment, registration, post-market monitoring, transparency, human oversight |
| Limited-risk | Chatbots, deepfakes, emotion recognition | Transparency: user must know they're interacting with AI |
| Minimal-risk | Recommendation systems, spam filters, most B2B SaaS internals | No specific obligations |
Run ai_risk_classifier.py to classify a use case and get the required-controls list.
US state patchwork (non-exhaustive):
Industry-specific overlays:
See references/ai_risk_governance.md for the full regulatory landscape + governance program checklist.
The breakeven question: at what monthly token volume does self-hosted inference beat API costs?
Key components:
Typical breakeven (frontier-quality): 100M–500M tokens/month, depending on model size and acceptable quality tradeoff. Below this, API wins. Above this, run the calculator.
Run ai_cost_economics.py with workload characteristics for a breakeven point + sensitivity to GPU rates and model size.
See references/ai_cost_economics.md for the full economics model and operational considerations.
The wrong question: "Should we hire an ML engineer or a research scientist?" The right question: "What's the next AI capability we need to ship, and what role unblocks that?"
Stage-to-role map:
| Stage | First AI hire | Then | Then |
|---|---|---|---|
| Pre-PMF | Founder + 1 ML-curious engineer playing with prompts | — | — |
| Series A | AI engineer (applied, full-stack; owns prompts/evals/deployment) | Second AI engineer for evals/quality | — |
| Series B | AI/ML platform engineer (inference, evals, observability) | Third AI engineer for production reliability | Data scientist if model is core IP |
| Series C | Manager of AI | ML research scientist (only if model IS the product) | AI safety / red team (if customer-facing AI) |
| Late-stage | Head of AI → CAIO | Multiple research scientists, platform team, safety/red team | Federated AI leads per business unit |
Critical distinctions:
Centralize-vs-embed for AI: AI starts centralized (one team) and stays there longer than data team, because the surface area is smaller. Embed only when AI is being deployed in 4+ product surfaces.
See references/ai_team_org_evolution.md.
Goal: Decide whether a specific use case should use API, fine-tune, or build.
# 1. Define use_case.json (volume, latency, accuracy, team size, budget)
python scripts/model_buildvsbuy_calculator.py use_case.json
# 2. Review 3-year TCO + breakeven
# 3. Cross-check with cs-cfo-advisor on budget commitment
# 4. Cross-check with cs-cto-advisor on engineering capacity (esp. for fine-tune)
# 5. Log via /cs:decide; consider /cs:freeze 60 on multi-year vendor commitment
Goal: Classify a use case under EU AI Act + US state laws, identify required controls.
# 1. Define use_case.json (decisions affected, users, geography, sector)
python scripts/ai_risk_classifier.py use_case.json
# 2. For HIGH-RISK: budget conformity assessment + registration
# 3. For LIMITED-RISK: implement transparency requirements
# 4. Cross-check with cs-general-counsel-advisor on contractual implications
# 5. Cross-check with cs-ciso-advisor on technical safeguards
# 6. Log via /cs:decide
Goal: Decide when (and whether) to migrate from API to self-hosted inference.
# 1. Build workload.json (tokens/day, model size, latency, quality tolerance)
python scripts/ai_cost_economics.py workload.json
# 2. Run sensitivity scenarios (low/mid/high GPU rates)
# 3. Estimate migration cost (engineering time + risk)
# 4. Cross-check with cs-cfo-advisor on capex commitment
# 5. Cross-check with cs-cto-advisor on platform readiness
# 6. Log via /cs:decide; pair with /cs:freeze if signing GPU commitment
Goal: Sequence next 18 months of AI hires aligned to capabilities to ship.
ai_team_org_evolution.md)**Bottom Line:** [one sentence — decision and rationale]
**The Decision:** [one of: model selection | risk classification | economics | next hire]
**The Evidence:** [numbers from the tool, not adjectives]
**How to Act:** [3 concrete next steps]
**Your Decision:** [the call only the founder can make]
../chief-data-officer-advisor/ — Training data rights, data product strategy (chains directly to model decisions)../cto-advisor/ — Architecture capacity, scaling cliffs (esp. for self-hosted inference)../ciso-advisor/ — Threat modeling for AI (prompt injection, jailbreak, training data poisoning)../general-counsel-advisor/ — AI contracts (vendor liability, output ownership, training-data licensing)../cfo-advisor/ — Build-vs-buy TCO math, multi-year vendor commitments../chro-advisor/ — AI team hiring + comp../../../engineering/rag-architect/ — Tactical RAG implementation../../../engineering/agent-designer/ — Tactical agent architecture../../../engineering/prompt-governance/ — Tactical prompt management../../../engineering/self-eval/ — Tactical eval infrastructure../../../engineering/llm-cost-optimizer/ — Tactical inference cost optimizationVersion: 1.0.0 Status: Production Ready Disclaimer: AI regulation is evolving rapidly. This skill surfaces decisions and tradeoffs as of 2026 but cannot replace qualified AI counsel for binding compliance decisions, especially under EU AI Act conformity assessments.