šØ EXECUTION NOTICE FOR CLAUDE
When you invoke this command via SlashCommand, the system returns THESE INSTRUCTIONS below.
YOU are the executor. This is NOT an autonomous subprocess.
- ā
The phases below are YOUR execution checklist
- ā
YOU must run each phase immediately using tools (Bash, Read, Write, Edit, TodoWrite)
- ā
Complete ALL phases before considering this command done
- ā DON't wait for "the command to complete" - YOU complete it by executing the phases
- ā DON't treat this as status output - it IS your instruction set
Immediately after SlashCommand returns, start executing Phase 0, then Phase 1, etc.
See @CLAUDE.md section "SlashCommand Execution - YOU Are The Executor" for detailed explanation.
Security Requirements
CRITICAL: All generated files must follow security rules:
@docs/security/SECURITY-RULES.md
Key requirements:
- Never hardcode API keys or secrets
- Use placeholders:
your_service_key_here
- Protect
.env files with .gitignore
- Create
.env.example with placeholders only
- Document key acquisition for users
Arguments: $ARGUMENTS
Goal: Monitor active ML training jobs on cloud platforms (Modal, AWS, GCP) and display real-time metrics including job status, GPU utilization, training progress, estimated completion time, and cost
Core Principles:
- Detect cloud platform from job ID format or configuration
- Query platform API for real-time job status
- Display metrics in clear, actionable format
- Calculate cost estimates based on GPU hours
Available Skills
This commands has access to the following skills from the ml-training plugin:
- cloud-gpu-configs: Platform-specific configuration templates for Modal, Lambda Labs, and RunPod with GPU selection guides
- cost-calculator: Cost estimation scripts and tools for calculating GPU hours, training costs, and inference pricing across Modal, Lambda Labs, and RunPod platforms. Use when estimating ML training costs, comparing platform pricing, calculating GPU hours, budgeting for ML projects, or when user mentions cost estimation, pricing comparison, GPU budgeting, training cost analysis, or inference cost optimization.
- example-projects: Provides three production-ready ML training examples (sentiment classification, text generation, RedAI trade classifier) with complete training scripts, deployment configs, and datasets. Use when user needs example projects, reference implementations, starter templates, or wants to see working code for sentiment analysis, text generation, or financial trade classification.
- integration-helpers: Integration templates for FastAPI endpoints, Next.js UI components, and Supabase schemas for ML model deployment. Use when deploying ML models, creating inference APIs, building ML prediction UIs, designing ML database schemas, integrating trained models with applications, or when user mentions FastAPI ML endpoints, prediction forms, model serving, ML API deployment, inference integration, or production ML deployment.
- monitoring-dashboard: Training monitoring dashboard setup with TensorBoard and Weights & Biases (WandB) including real-time metrics tracking, experiment comparison, hyperparameter visualization, and integration patterns. Use when setting up training monitoring, tracking experiments, visualizing metrics, comparing model runs, or when user mentions TensorBoard, WandB, training metrics, experiment tracking, or monitoring dashboard.
- training-patterns: Templates and patterns for common ML training scenarios including text classification, text generation, fine-tuning, and PEFT/LoRA. Provides ready-to-use training configurations, dataset preparation scripts, and complete training pipelines. Use when building ML training pipelines, fine-tuning models, implementing classification or generation tasks, setting up PEFT/LoRA training, or when user mentions model training, fine-tuning, classification, generation, or parameter-efficient tuning.
- validation-scripts: Data validation and pipeline testing utilities for ML training projects. Validates datasets, model checkpoints, training pipelines, and dependencies. Use when validating training data, checking model outputs, testing ML pipelines, verifying dependencies, debugging training failures, or ensuring data quality before training.
To use a skill:
!{skill skill-name}
Use skills when you need:
- Domain-specific templates and examples
- Validation scripts and automation
- Best practices and patterns
- Configuration generators
Skills provide pre-built resources to accelerate your work.
Phase 1: Job Discovery
Goal: Parse job ID and identify cloud platform
Actions:
- Parse $ARGUMENTS for job ID
- If no job ID provided, list all active jobs across platforms
- Detect platform from job ID format:
- Modal: starts with "ap-" or use modal CLI
- AWS SageMaker: job-* format
- GCP Vertex AI: projects/* format
- Check if Modal CLI available: !{bash which modal}
- Load .env for API credentials if needed: !{bash test -f .env && echo "found" || echo "missing"}
Phase 2: Query Job Status
Goal: Retrieve job information from cloud platform
Actions:
- Modal Platform:
- List running apps: !{bash modal app list --json}
- Get specific job status: !{bash modal app logs $ARGUMENTS --raw}
- Parse JSON output for job metadata
- AWS SageMaker (if detected):
- Query job: !{bash aws sagemaker describe-training-job --training-job-name $ARGUMENTS}
- Get CloudWatch metrics for GPU utilization
- GCP Vertex AI (if detected):
- Query job: !{bash gcloud ai custom-jobs describe $ARGUMENTS}
- Get resource metrics
- If platform not detected, report: "Unable to detect platform from job ID: $ARGUMENTS"
Phase 3: Metrics Display
Goal: Format and display training metrics
Actions:
Phase 4: Cost Calculation
Goal: Estimate training costs based on GPU usage
Actions:
Phase 5: Summary
Goal: Display actionable information and next steps
Actions:
- Show summary status:
- Job {job-id} is {Running|Completed|Failed}
- {elapsed} elapsed, {eta} remaining
- Current cost: ${amount}
- If job is running:
- "Use 'modal app logs {job-id}' to stream full logs"
- "Use 'modal app stop {job-id}' to terminate job"
- If job completed:
- "Check models/checkpoints/ for saved model"
- "Review logs/training/ for metrics history"
- If job failed:
- "Review error logs: modal app logs {job-id} --tail 50"
- "Check GPU memory errors or timeout issues"
- Refresh suggestion: "Re-run this command to update metrics"