From posthog
Diagnoses failed data warehouse syncs from Stripe, Postgres, Hubspot sources; handles stuck runs, credentials, schema drift, incremental misconfigs; recommends cancel, reload, resync, delete-data.
npx claudepluginhub anthropics/claude-plugins-official --plugin posthogThis skill uses the workspace's default tool permissions.
Work top-down when a data warehouse source or table is failing, stuck, or producing bad data: source → schema →
Guides connecting data warehouse sources like Postgres, MySQL, Stripe, HubSpot, MongoDB, Salesforce, BigQuery, Snowflake to PostHog via wizard, credential validation, table discovery, sync selection, and creation.
Diagnoses and resolves errors in SAP Datasphere Data Flows, Replication Flows, Transformation Flows, and Task Chains. Guides troubleshooting failed executions, partial loads, performance issues, and connectivity problems.
Migrates Reverse ETL syncs from Census, Hightouch, Polytomic, or custom scripts to drt by mapping sources, destinations, schedules, and generating sync YAML configs.
Share bugs, ideas, or general feedback.
Work top-down when a data warehouse source or table is failing, stuck, or producing bad data: source → schema → recovery action. Do not jump straight to "resync from scratch" — that discards synced data and restarts from zero, which is rarely the right first step.
Running state far longer than expectedCompletedauditing-warehouse-data-health — has surfaced a failing source or schema and the user
wants to dig into itBoth entry points (user-reported and audit-handoff) use the same workflow; the audit just means you already know which item to diagnose and can skip Step 1's discovery search.
| Tool | Purpose |
|---|---|
external-data-sources-list | List all sources with connection status and latest error |
external-data-sources-retrieve | Full details for one source including all its schemas |
external-data-schemas-list | All table schemas across all sources, with per-table status + latest_error |
external-data-schemas-retrieve | Full details for one schema including sync_type_config |
external-data-schemas-cancel | Cancel a sync currently in Running state |
external-data-schemas-reload | Trigger a sync using the configured sync method (respects incremental) |
external-data-schemas-resync | Full resync — wipes synced data and restarts. Destructive |
external-data-schemas-delete-data | Delete the synced table but keep the schema entry |
external-data-schemas-partial-update | Change sync_type / incremental_field / cdc_table_mode |
external-data-sources-partial-update | Update a source's credentials (job_inputs) after rotation |
external-data-sources-reload | Retrigger syncs for every enabled schema on a source |
external-data-sources-refresh-schemas | Re-fetch the source's table list to pick up new tables |
external-data-sources-check-cdc-prerequisites-create | Verify Postgres CDC setup for a source |
external-data-schemas-incremental-fields-create | Refresh candidate incremental fields when the source schema has changed |
external-data-sources-webhook-info-retrieve | Check webhook registration state and external service status |
external-data-sources-create-webhook-create | Re-register a webhook that was lost or never registered |
external-data-sources-update-webhook-inputs-create | Update the signing secret after rotation on the source side |
external-data-sources-delete-webhook-create | Remove a broken webhook before re-registering |
If the user named a source, go straight to external-data-sources-retrieve. Otherwise start with
external-data-sources-list and external-data-schemas-list to find what's red.
Two kinds of failure:
ExternalDataSource.status = "Error"): the connection itself is broken — credentials expired,
host unreachable, account disabled. Affects every table.external-data-schemas-list, look for status values "Failed", "Billing limits", or "Billing limits too low". (The underlying model enum values are BillingLimitReached and BillingLimitTooLow, but the
serializer rewrites them — match on both the human-readable and enum forms to be safe.)A source can look Completed at the top level while one of its schemas is Failed — always check both.
From external-data-schemas-list, each schema has a status:
| Status | Meaning | Usually means |
|---|---|---|
Running | Sync currently executing | Normal, unless stuck for hours |
Completed | Last sync finished successfully | Healthy |
Failed | Last sync errored — see latest_error | Needs diagnosis |
Paused | User disabled sync (should_sync = false) | Intentional |
Billing limits (serializer) / BillingLimitReached (enum) | Team hit its warehouse row quota | Billing issue, not a technical failure |
Billing limits too low (serializer) / BillingLimitTooLow (enum) | Team has insufficient credit | Billing issue |
Always check last_synced_at alongside status. A schema in Running with last_synced_at from 12 hours ago is
almost certainly stuck, even though the status isn't Failed.
latest_errorMap the latest_error string to a root cause. Common patterns:
| Error substring | Root cause | Fix |
|---|---|---|
authentication failed, 401, 403, invalid credentials | Credentials expired or rotated | User rotates creds, then external-data-sources-partial-update with new job_inputs |
Could not establish session to SSH gateway | SSH tunnel misconfigured or remote host down | User checks SSH host/key/bastion |
Primary key required for incremental syncs | Table has no PK and sync_type is incremental/cdc | Either add PK in source, or switch schema to full_refresh |
primary keys for this table are not unique | Declared PK columns aren't actually unique | Pick different PK columns via partial-update |
Integration matching query does not exist | Source's saved integration was deleted | Recreate the source |
column "X" does not exist, does not have a column named | Schema drift — incremental field or tracked column removed | Use incremental-fields-create to re-detect, then partial-update |
relation "..." does not exist | Source table was dropped/renamed | Remove schema or rename source-side |
SSL, connection refused, timeout, unreachable | Network / firewall / host reachability | User side — check host/port/allowlist |
replication slot, publication, wal_level | CDC prerequisites broken | Run check-cdc-prerequisites-create; may need slot recreate |
Schema exceeds row limit, billing | Billing limit | Upgrade plan or disable the schema |
If latest_error is null but the schema is Failed, retrieve the schema directly — the error may only be populated
on the detail view.
The recovery action depends on root cause, not just status. Match the user's situation to one of these:
A. Transient failure (network blip, temporary API outage)
external-data-schemas-reload to retry using the configured sync method.B. Credentials expired or rotated
external-data-sources-partial-update with the new job_inputs → the reload happens
automatically when the source status flips back to running, or trigger manually with external-data-sources-reload.C. Schema drift — column renamed, dropped, or type changed
external-data-schemas-incremental-fields-create to get the current fields, then
external-data-schemas-partial-update with the corrected incremental_field / incremental_field_type /
primary_key_columns. Usually no need to wipe data.C2. Added / renamed tables in the source database
external-data-sources-refresh-schemas to pick up the new table list, then configure sync on any new
schemas.D. Incremental state is wrong (duplicates, missing rows, data looks corrupt)
Completed — this isn't a "failure" per se, it's bad data.external-data-schemas-resync to wipe synced data and re-import from source. Destructive but often the
right call for data-quality issues.E. CDC pipeline broken on Postgres
external-data-sources-check-cdc-prerequisites-create to enumerate what's broken, fix on the Postgres
side, then external-data-schemas-reload. If the WAL position was lost, a resync is sometimes unavoidable.F. Sync is stuck in Running for hours
last_synced_at. If it's hours old and still Running, the job is orphaned.external-data-schemas-cancel to stop it, then external-data-schemas-reload.G. Table data is corrupt but you want to keep the schema config
external-data-schemas-delete-data to drop the synced table but preserve the schema entry. Next reload
re-imports from scratch without losing the configured sync_type/incremental_field.H. Billing limit
I. Webhook-backed schema isn't receiving events
sync_type: "webhook", initial bulk sync finished, but no new rows arrive despite activity on
the source side. Status may still read Completed because the bulk sync (the safety-net cadence) is succeeding —
the problem is the push path.external-data-sources-webhook-info-retrieve({source_id}).exists: false → the webhook was never registered, or was deleted. Call create-webhook-create to
register it.exists: true but external_status.error is set → typically "API key doesn't have permission to read
webhooks" or similar. The webhook may have been deleted on the source's dashboard. Re-create it.external_status.status isn't "enabled" → the source disabled the webhook (usually after repeated
delivery failures). Re-enable or re-register.update-webhook-inputs-create({source_id}, {inputs: {signing_secret: "..."}}).Three recovery actions discard data and cannot be undone:
external-data-schemas-resync — wipes synced rows, re-imports from scratchexternal-data-schemas-delete-data — drops the synced tableexternal-data-sources-destroy — deletes the source and all its schemasAlways present the fix you're proposing and wait for explicit approval before calling any of these. "Just try resync" is rarely the right default.
User: "Our Stripe sync is broken, can you check?"
Agent:
- external-data-sources-list → find Stripe source, status = Error
- external-data-sources-retrieve({id}) → latest_error: "authentication failed: 401 Unauthorized"
- Report: "Your Stripe source's API key is no longer authenticating.
All 8 tables under it are failing with 401s. This usually means the key was rotated on the Stripe side.
To fix:
1. Grab a fresh restricted API key from the Stripe dashboard.
2. I'll update the source with the new key.
3. Syncs will resume automatically — no data loss.
Paste the new key here when ready."
User: "sk_live_..."
Agent:
- external-data-sources-partial-update({id}, {job_inputs: {stripe_secret_key: "sk_live_..."}})
- external-data-sources-reload({id}) to trigger retry
- Report: "Updated and re-triggered. Check back in a few minutes — latest_error should clear."
Error, nothing under it will work;
fixing the source usually fixes all its schemas at once.Running isn't always healthy. Cross-check last_synced_at. A sync stuck in Running needs cancel then
reload, not resync.delete-data + reload over resync + new schema entry — it keeps the configured sync_type / incremental_field
/ PK setup.Completed even when the push channel is broken. When users say "my data is hours behind" on a webhook schema,
call webhook-info-retrieve before looking at schema status. Webhook issues don't surface on
external-data-schemas-list.