Implement Vercel load testing, auto-scaling, and capacity planning strategies. Use when running performance tests, configuring horizontal scaling, or planning capacity for Vercel integrations. Trigger with phrases like "vercel load test", "vercel scale", "vercel performance test", "vercel capacity", "vercel k6", "vercel benchmark".
From vercel-packnpx claudepluginhub nickloveinvesting/nick-love-plugins --plugin vercel-packThis skill is limited to using the following tools:
references/capacity-planning.mdreferences/errors.mdreferences/examples.mdreferences/load-testing-with-k6.mdreferences/scaling-patterns.mdGuides Next.js Cache Components and Partial Prerendering (PPR) with cacheComponents enabled. Implements 'use cache', cacheLife(), cacheTag(), revalidateTag(), static/dynamic optimization, and cache debugging.
Migrates code, prompts, and API calls from Claude Sonnet 4.0/4.5 or Opus 4.1 to Opus 4.5, updating model strings on Anthropic, AWS, GCP, Azure platforms.
Details PluginEval's skill quality evaluation: 3 layers (static, LLM judge), 10 dimensions, rubrics, formulas, anti-patterns, badges. Use to interpret scores, improve triggering, calibrate thresholds.
Load testing Vercel deployments requires understanding the platform's auto-scaling model: serverless functions scale horizontally without configuration, but they have per-invocation concurrency limits and cold-start penalties that can cause latency spikes under sudden traffic bursts. Load tests should simulate both sustained load (to measure steady-state performance) and spike scenarios (to measure cold-start amplification) to give a complete picture of behavior under production traffic patterns.
Write a k6 test script that models your realistic traffic pattern including the read/write ratio, request size distribution, and geographic origin if applicable. Set thresholds for P95 and P99 latency at values aligned to your user experience SLAs. Include a warm-up phase to establish a baseline before the peak load phase to isolate cold-start latency from steady-state latency.
Vercel serverless functions scale automatically, but you may need to configure concurrency limits or use Vercel's Edge Functions for lower-latency scaling at the edge. For workloads with predictable spikes, consider pre-warming strategies or migrating latency-critical paths to edge functions which eliminate cold starts entirely.
Execute the test against a staging deployment that mirrors production configuration. Collect function invocation metrics from the Vercel dashboard alongside your k6 output. Watch for error rate increases at high concurrency levels and log cold-start occurrences to quantify their frequency and duration.
Record results in the benchmark template documenting throughput, latency percentiles, cold-start frequency, and error rate at each load level. Derive capacity recommendations identifying the traffic volume at which latency SLAs are breached.
See ${CLAUDE_SKILL_DIR}/references/errors.md for comprehensive error handling.
See ${CLAUDE_SKILL_DIR}/references/examples.md for detailed examples.
Implement Vercel load testing, auto-scaling, and capacity planning strategies.