From rest-api-pipeline
Finds dlt sources and connectors for APIs, databases, or files by classifying requests, searching verified sources, and recommending init commands for data pipelines.
npx claudepluginhub dlt-hub/dlthub-ai-workbench --plugin rest-api-pipelineThis skill uses the workspace's default tool permissions.
Locate the best dlt source for what the user wants to extract data from.
Scaffolds minimal dlt REST API pipeline via dlt init command for rest_api core source or generic HTTP APIs. Excludes sql_database/filesystem sources.
Plans DataHub connectors by classifying source systems, researching via agent or inline, and generating _PLANNING.md blueprints with entity mappings and architecture decisions. For new connector design or source research.
Operates anysite CLI for web data extraction from LinkedIn/Instagram/Twitter, batch API processing, dataset pipelines with scheduling/transforms/exports, SQL queries, PostgreSQL/SQLite loading, and LLM data analysis (summarize/classify/enrich).
Share bugs, ideas, or general feedback.
Locate the best dlt source for what the user wants to extract data from.
Parse $ARGUMENTS:
source-name (required): what the user wants to extract data from (e.g., "alpaca markets", "stripe", "postgres", "csv files", "rest api")| User says (examples) | Core source |
|---|---|
| postgres, mysql, mssql, oracle, database, db, sql | sql_database |
| rest api, http api, web api, rest | rest_api |
| files, csv, parquet, jsonl, s3, gcs, azure blob, local files | filesystem |
If it matches a core source, skip to step 5 and report the core source match.
If the request looks like a specific API/service name, run:
dlt --non-interactive init --list-sources
Search the output (case-insensitive) for the source name. If found, ensure that the verified source contains the data that the user needs (ask the user explicitly) skip to step 5
Use search_dlthub_sources mcp tool to look for sources. It is full-text search based so pass only essential keywords to it
ie. "claude analytics". You'll get description of the source and set of reference links to use in web search below.
query: <source-name> API documentation
NOTE: we can handle only REST API (step 5) and sometimes GraphQL.
This toolkit builds REST API pipelines. Before continuing, check if the user's data source actually fits.
STOP and hand off if any of these are true:
rest_api — the user needs sql_database, filesystem, or another core source. Tell them which one and the dlt init command, then suggest a general coding session to build the pipeline.dlt init <source> <destination> command. Suggest they try the verified source first.Found: <verified source or non-REST core source>
Command: dlt init <source> <destination>
This is outside the REST API pipeline workflow. You can:
1. Use the verified source / core source above (recommended)
2. Start a general coding session if you need a custom pipeline
CONTINUE only when the best path is building a REST API pipeline — either because:
rest_api core source under the hood)Ask user to pick a single endpoint to start the work - do it directly or infer it from conversation.
Do NOT run dlt init yet — wait for user confirmation.
After that continue workflow in create-rest-api-pipeline skill