From amplitude
Triages product reliability issues by investigating Amplitude events for network failures, JavaScript errors, and error clicks to pinpoint what's broken, affected users, and causes.
npx claudepluginhub amplitude/mcp-marketplace --plugin amplitudeThis skill uses the workspace's default tool permissions.
Investigate product errors by triaging across three auto-captured event types — `[Amplitude] Network Request`, `[Amplitude] Error Logged`, and `[Amplitude] Error Click` — to identify what's broken, which users are affected, and what's causing it. This skill cross-references all three signals to surface causal chains (failed request → JS error → user frustration) rather than treating each in iso...
Guides Payload CMS config (payload.config.ts), collections, fields, hooks, access control, APIs. Debugs validation errors, security, relationships, queries, transactions, hook behavior.
Builds scalable data pipelines, modern data warehouses, and real-time streaming architectures using Spark, dbt, Airflow, Kafka, and cloud platforms like Snowflake, BigQuery.
Builds production Apache Airflow DAGs with best practices for operators, sensors, testing, and deployment. For data pipelines, workflow orchestration, and batch job scheduling.
Investigate product errors by triaging across three auto-captured event types — [Amplitude] Network Request, [Amplitude] Error Logged, and [Amplitude] Error Click — to identify what's broken, which users are affected, and what's causing it. This skill cross-references all three signals to surface causal chains (failed request → JS error → user frustration) rather than treating each in isolation.
This is a reactive investigation skill — the user has a signal (spike, complaint, experiment regression, gut feeling) and wants to understand what's happening. For proactive monitoring, use the monitor-reliability skill instead.
These are the three auto-captured events this skill operates on. Never guess property names — use exactly these.
[Amplitude] Network Request — Browser network requests.
Key properties: [Amplitude] URL, [Amplitude] Status Code, [Amplitude] Duration, [Amplitude] Request Method, [Amplitude] Request Type, [Amplitude] Request Body Size, [Amplitude] Response Body Size, [Amplitude] Start Time, [Amplitude] Completion Time, [Amplitude] Page Path.
[Amplitude] Error Logged — JavaScript errors.
Key properties: Error Message, Error Type, Error URL, File Name, Error Lineno, Error Colno, Error Stack Trace, [Amplitude] Error Detection Source.
[Amplitude] Error Click — Clicks on error-associated UI elements.
Key properties: [Amplitude] Message, [Amplitude] Kind, [Amplitude] Filename, [Amplitude] Line Number, [Amplitude] Column Number, [Amplitude] Element Text, [Amplitude] Element Tag, [Amplitude] Element Hierarchy.
All three share: [Amplitude] Page Path, [Amplitude] Page URL, [Amplitude] Session Replay ID.
Amplitude:get_context. If multiple projects, ask which to investigate.[Amplitude] Network Request, then check if they cascade into JS errors[Amplitude] Error Logged, filtered to that errorRun these in parallel where possible. Budget: 4-6 calls for this step.
Use Amplitude:query_dataset to query [Amplitude] Network Request:
[Amplitude] Status Code to 4xx and 5xx ranges. Measure daily event counts and unique users. Compare to total network request volume for a failure rate percentage.[Amplitude] URL to rank which APIs fail most. Include [Amplitude] Status Code as a secondary grouping to distinguish 401s (auth) from 500s (server errors) from 404s (missing).[Amplitude] Duration by [Amplitude] URL. Flag P95 > 3s or mean > 1s.Use Amplitude:query_dataset to query [Amplitude] Error Logged:
Error Message to find the highest-volume errors. Include Error Type and File Name for context.Use Amplitude:query_dataset to query [Amplitude] Error Click:
[Amplitude] Element Text or [Amplitude] Message to see which error UI elements get the most interaction.This is where the skill adds value beyond looking at each event in isolation.
Failed request → JS error chain. Compare the timing and pages of network failures (Step 2a) with JS errors (Step 2b). If the same pages have both 5xx network failures AND JS errors, the network failure is likely the root cause. Use [Amplitude] Page Path as the join dimension.
Error → frustration chain. Compare JS error pages with error click pages. High error click volume on pages with high JS error rates confirms users are seeing and interacting with the broken experience.
Page-level triage. Use Amplitude:query_dataset to group all three events by [Amplitude] Page Path. Produce a page-level error heatmap:
For the top 2-3 error patterns from Step 3:
Amplitude:query_dataset to count unique users affected. Compare to total active users for an impact percentage.Amplitude:get_event_properties if you need to discover available properties.Amplitude:get_session_replays filtered to sessions containing the error event. Provide 2-3 replay links so the user can see exactly what happened.Build a root cause hypothesis using evidence from the prior steps:
Amplitude:get_deployments once. Check if error spikes align with recent deploys. If a deployment shipped within 24 hours of the error spike, it's the leading hypothesis.Amplitude:get_experiments and Amplitude:query_experiment to check.Amplitude:get_feedback_sources then Amplitude:get_feedback_insights with keywords from the top error messages. If users are reporting the same issue, it validates the impact and may provide additional context the data can't.Structure the output as a triage report. Lead with what's most broken and actionable.
Required sections:
Diagnosis summary (2-3 sentences): The single most important finding. Written as a headline you'd send to the engineering lead. Include scope: how many users, which pages, since when.
Error landscape — A table summarizing the state across all three signal types:
| Signal | Volume (7d) | Trend | Top Source | Severity |
|--------|-------------|-------|------------|----------|
| Network failures (4xx/5xx) | [N] requests | [↑/↓/→] | [endpoint] | [Critical/High/Medium/Low] |
| JS errors | [N] errors, [N] users | [↑/↓/→] | [Error Message] | ... |
| Error clicks | [N] clicks | [↑/↓/→] | [Element Text] | ... |
Top errors (3-5 max): Each as a narrative paragraph:
Causal chains (if found): Describe the cross-event chain. "POST to /api/query is returning 500 → this triggers an unhandled TypeError in ChartRenderer.tsx:142 → users see and click the error state. ~1,200 users affected in the last 7 days."
Recommended actions (2-4 numbered items): Concrete, copy-paste-ready. Start each with a verb. Bias toward fixing, investigating further with a specific breakdown, or setting up monitoring.
Follow-on prompt: Ask what to dig into next — e.g., "Want me to segment the API failures by org tier, watch a few session replays, or build a monitoring dashboard for these errors?"
Severity classification:
| Severity | Criteria |
|---|---|
| Critical | >5% of users affected, full causal chain (network → error → frustration), or blocking a core flow |
| High | 1-5% of users, JS errors on key pages, or a clear regression from a deploy |
| Medium | <1% of users, chronic errors, or errors on non-critical pages |
| Low | Silent errors with no user-facing impact, or errors isolated to a single edge-case segment |
[Amplitude] Network Request, [Amplitude] Error Logged, or [Amplitude] Error Click events. These require Session Replay or the autocapture plugin to be enabled." Suggest the user check their SDK configuration.[Amplitude] Error Logged by Error Message. Then check for correlated network failures and error clicks.User says: "What's broken right now?"
Actions:
User says: "Errors seem up since yesterday's deploy"
Actions:
get_deployments for what shipped[Amplitude] Error Logged comparing pre-deploy (7d before) vs post-deploy (last 24h)User says: "We're seeing a lot of TypeErrors in the chart builder"
Actions:
[Amplitude] Error Logged to Error Type = TypeError and [Amplitude] Page Path containing the chart builderError Message and File Name to find the specific errors[Amplitude] Network Request on the same pages for failing API calls