From amplitude
Analyzes user queries to AI agents via Amplitude Agent Analytics, identifying top topics, quality scores, failures, coverage gaps, and product improvement opportunities.
npx claudepluginhub amplitude/mcp-marketplace --plugin amplitudeThis skill uses the workspace's default tool permissions.
You analyze what users ask AI agents about and how well each topic is served — surfacing underserved areas, coverage gaps, and product opportunities from conversation patterns. This is the product intelligence skill that turns AI session data into "what to build next" decisions.
Guides Payload CMS config (payload.config.ts), collections, fields, hooks, access control, APIs. Debugs validation errors, security, relationships, queries, transactions, hook behavior.
Builds scalable data pipelines, modern data warehouses, and real-time streaming architectures using Spark, dbt, Airflow, Kafka, and cloud platforms like Snowflake, BigQuery.
Builds production Apache Airflow DAGs with best practices for operators, sensors, testing, and deployment. For data pipelines, workflow orchestration, and batch job scheduling.
You analyze what users ask AI agents about and how well each topic is served — surfacing underserved areas, coverage gaps, and product opportunities from conversation patterns. This is the product intelligence skill that turns AI session data into "what to build next" decisions.
Amplitude:get_context to identify projects and user role.Amplitude:get_agent_analytics_schema with include: ["filter_options", "taxonomy"] to discover available topic models, agent names, and classification values. The schema tells you what topic dimensions exist (e.g., product_area, intent, error_domain) — these vary by project.Run these in parallel:
Topic breakdown with quality. Call Amplitude:query_agent_analytics_metrics with metrics: ["topics"], limit: 50. This returns each topic with session count, average quality score, average sentiment, and failure rate. This is the core dataset.
Agent-by-topic matrix. Call Amplitude:query_agent_analytics_sessions with groupBy: ["agent_name", "primary_topic"], limit: 100. This shows which agents handle which topics — and where quality differs by agent for the same topic.
Volume trend by topic. Call Amplitude:query_agent_analytics_metrics with metrics: ["volume_timeseries"], interval: "DAY". While this is aggregate, combine it with the topic breakdown to understand whether total volume growth is driven by specific topics.
Failure sessions by topic. Call Amplitude:query_agent_analytics_sessions with hasTaskFailure: true, groupBy: ["primary_topic"], limit: 50. This shows which topics have the most failures — a different signal from low quality (failures are hard stops, low quality is soft degradation).
Score each topic on a 2x2 of volume x quality:
| High Quality (>0.7) | Low Quality (<0.7) | |
|---|---|---|
| High Volume | Well-served (maintain) | Underserved (fix now) |
| Low Volume | Niche but working (monitor) | Gap or emerging (investigate) |
For each quadrant, identify the top 3-5 topics. The high volume + low quality quadrant is the priority — these are things users frequently ask about that the agents handle poorly.
Also flag:
For the 2-3 most impactful underserved topics:
Sample conversations. Call Amplitude:search_agent_analytics_conversations with keywords from the topic to find representative conversations. Read 3-5 examples to understand:
Detailed failing sessions. Call Amplitude:query_agent_analytics_sessions filtered to the topic with hasTaskFailure: true or maxQualityScore: 0.5, responseFormat: "detailed", limit: 5. Read the enrichment data for failure reasons and rubric scores.
Tool usage for the topic. Call Amplitude:query_agent_analytics_metrics with metrics: ["tool_stats"] — if you can filter to sessions for this topic. Otherwise, pull span data for a few failing sessions with Amplitude:query_agent_analytics_spans to see which tools are involved.
Transform the analysis into actionable product decisions.
Required sections:
Topic landscape summary (3-4 sentences): How many distinct topics, total session volume, overall quality distribution. Frame as "your AI agents handle X topics across Y sessions — here's what's working and what isn't."
Topic heatmap table — The core deliverable:
| Topic | Sessions | Quality | Sentiment | Failure Rate | Trend | Priority |
|-------|----------|---------|-----------|--------------|-------|----------|
| [topic] | [N] | [score] | [score] | [%] | [↑/↓/→] | [Fix/Monitor/Good] |
Sort by priority: Fix items first, then Monitor, then Good. Limit to top 15-20 topics.
Underserved topics (2-4 findings): Each as a narrative paragraph:
Coverage gaps (1-2 findings): Topics where users are asking questions the agents can't answer at all. Evidence: high failure rates, very low quality, or sessions where the agent explicitly says "I can't help with that."
Emerging topics (1-2 findings): Topics with growing volume that may need attention soon. Include the growth rate and current quality.
Agent routing insights (if applicable): Topics that would be better served by a different agent, or topics where adding a specialized agent would improve quality.
Recommended actions (3-5 numbered items): Prioritized by impact (volume x quality gap). Examples:
Follow-on prompt: "Want me to deep-dive into a specific topic, investigate the failing sessions for [top underserved topic], or build a monitoring dashboard for AI topic quality?"
Writing standards:
User says: "What are people asking our AI about?"
Actions:
User says: "Where is our AI falling short?"
Actions:
User says: "What topics does the Chart Agent handle, and how well?"
Actions:
Topics require session enrichment to be enabled. If topics are empty, fall back to search_agent_analytics_conversations with broad keyword searches to manually categorize common themes. Note the limitation and suggest enabling enrichment.
Group similar topics and present the top 20 by volume. Offer to drill into specific clusters on request.
If topic labels are broad (e.g., "data question", "help request"), the enrichment model may need tuning. Note this and use conversation search to identify more specific sub-topics manually.