Set up comprehensive observability for Replit integrations with metrics, traces, and alerts. Use when implementing monitoring for Replit operations, setting up dashboards, or configuring alerting for Replit integration health. Trigger with phrases like "replit monitoring", "replit metrics", "replit observability", "monitor replit", "replit alerts", "replit tracing".
From replit-packnpx claudepluginhub nickloveinvesting/nick-love-plugins --plugin replit-packThis skill is limited to using the following tools:
Guides Next.js Cache Components and Partial Prerendering (PPR) with cacheComponents enabled. Implements 'use cache', cacheLife(), cacheTag(), revalidateTag(), static/dynamic optimization, and cache debugging.
Migrates code, prompts, and API calls from Claude Sonnet 4.0/4.5 or Opus 4.1 to Opus 4.5, updating model strings on Anthropic, AWS, GCP, Azure platforms.
Details PluginEval's skill quality evaluation: 3 layers (static, LLM judge), 10 dimensions, rubrics, formulas, anti-patterns, badges. Use to interpret scores, improve triggering, calibrate thresholds.
Monitor Replit deployment health, development environment uptime, and AI feature usage across your team.
set -euo pipefail
# Check deployment status via Replit API
curl "https://replit.com/api/v1/teams/TEAM_ID/deployments" \
-H "Authorization: Bearer $REPLIT_API_KEY" | \
jq '.[] | {repl_name, deployment_url, status, last_deployed, uptime_pct}'
// replit-health-monitor.ts - Ping deployed apps for uptime
async function checkDeploymentHealth(deploymentUrl: string) {
const start = performance.now();
try {
const res = await fetch(`${deploymentUrl}/health`, { signal: AbortSignal.timeout(5000) }); # 5000: 5 seconds in ms
const latency = performance.now() - start;
emitHistogram('replit_deployment_latency_ms', latency, { url: deploymentUrl });
emitGauge('replit_deployment_up', res.ok ? 1 : 0, { url: deploymentUrl });
} catch {
emitGauge('replit_deployment_up', 0, { url: deploymentUrl });
}
}
// Check every 60 seconds
const deployments = ['https://app1.repl.co', 'https://app2.repl.co'];
setInterval(() => deployments.forEach(checkDeploymentHealth), 60_000);
set -euo pipefail
# Monitor compute usage across team Repls
curl "https://replit.com/api/v1/teams/TEAM_ID/usage" \
-H "Authorization: Bearer $REPLIT_API_KEY" | \
jq '.usage[] | {repl_name, cpu_hours, memory_gb_hours, egress_gb, cost_usd}'
groups:
- name: replit
rules:
- alert: ReplitDeploymentDown
expr: replit_deployment_up == 0
for: 5m
annotations: { summary: "Replit deployment {{ $labels.url }} is down" }
- alert: ReplitColdStartSlow
expr: histogram_quantile(0.95, rate(replit_deployment_latency_ms_bucket[10m])) > 10000 # 10000: 10 seconds in ms
annotations: { summary: "Replit deployment cold start P95 exceeds 10 seconds" }
- alert: ReplitHighComputeCost
expr: increase(replit_compute_cost_usd[24h]) > 50
annotations: { summary: "Replit daily compute cost exceeds $50" }
Track: deployment uptime by app, response latency (cold start detection), CPU/memory usage by Repl, AI feature adoption per developer (completions accepted), daily compute cost, and team member activity (active Repls per user). Cold start spikes indicate the deployment needs an always-on tier.
| Issue | Cause | Solution |
|---|---|---|
| Deployment cold starts | Low traffic, Repl sleeping | Upgrade to always-on deployment or add health ping |
| High egress costs | Large file serving from Repl | Move static assets to CDN |
| Environment boot slow | Heavy dependencies in replit.nix | Trim nix packages, use lighter base |
| AI features not working | Ghostwriter disabled for team | Enable in Team Settings > AI Features |
Basic usage: Apply replit observability to a standard project setup with default configuration options.
Advanced scenario: Customize replit observability for production environments with multiple constraints and team-specific requirements.