From data-agent-kit-starter-pack
Discovers and inspects BigQuery Data Transfer Service (DTS) configurations to identify existing ingestion pipelines and extract datasource or transfer metadata. Useful when building, managing data pipelines or handling ingest/add data requests.
npx claudepluginhub gemini-cli-extensions/data-agent-kit-starter-pack --plugin data-agent-kit-starter-packThis skill uses the workspace's default tool permissions.
> [!IMPORTANT]
Implements structured self-debugging workflow for AI agent failures: capture errors, diagnose patterns like loops or context overflow, apply contained recoveries, and generate introspection reports.
Monitors deployed URLs for regressions in HTTP status, console errors, performance metrics, content, network, and APIs after deploys, merges, or upgrades.
Provides React and Next.js patterns for component composition, compound components, state management, data fetching, performance optimization, forms, routing, and accessible UIs.
[!IMPORTANT]
All new BigQuery Data Transfer Service (DTS) configurations MUST be provisioned through the gcp pipeline resource provisioning framework, which includes generating a
deployment.yaml.
- Do NOT use imperative CLI commands (e.g.,
bq mkorgcloud) to create or update configurations.- CLI commands are permitted only for discovery (listing/showing) and triggering manual runs.
This guide enables the discovery of existing ingestion resources and provides metadata related to ingestion when needed.
Before generating configurations, discover the actual values for the target project and region.
[!TIP] If
deployment.yamlalready exists in the repository root, prioritize extractingprojectandregionfrom the target environment configuration (e.g.,dev).
gcloud config get projectgcloud config get compute/region[!TIP] Use these commands to replace placeholders like
<PROJECT_ID>with actual values. Always remove associated comments that start with TODO once replaced.
Before assuming a new transfer is needed, check for existing ones in the target region.
List Transfers:
bq ls --transfer_config \
--transfer_location=<REGION> \
--project_id=<PROJECT_ID>
Analyze Existing Transfers:
Single Transfer Found:
bq ls --transfer_run --transfer_config=<RESOURCE_NAME>Multiple Transfers Found:
Disabled Transfers Found:
deployment.yaml file by setting the disabled
field to false for the specific transfer resource.No Transfers Found: Proceed to create new if needed.
If creating a new transfer, discover the required parameters using the REST API and validate them with the user.
[!TIP] If
<DATA_SOURCE_ID>is unknown, run the discovery script without<DATA_SOURCE_ID>argument to list available source IDs (e.g.,google_cloud_storage). It uses the derived project and location from Step 0.python3 scripts/bigquery_dts.py --project_id=<PROJECT_ID>
Run Discovery Script: Use the bigquery_dts.py script to inspect Data
Source parameters via the REST API.
# Passes the derived project and region to the script.
python3 scripts/bigquery_dts.py --project_id=<PROJECT_ID> <DATA_SOURCE_ID> <REGION>
[!IMPORTANT] Run this command every time a new transfer is being planned.
[!CAUTION] Mandatory User Questionnaire (CRITICAL):
definition.versionInfo in deployment.yaml and then you can proceed.Wait for User Response: You MUST NOT proceed until parameters are confirmed.
Retrieve the configuration details for the selected transfer.
bq show --format=prettyjson --transfer_config <RESOURCE_NAME>
After the transfer is deployed via the resource provisioning framework, you MUST ensure there is at least a single successful run before proceeding with the rest of the tasks.
Trigger a Manual Run: If no successful runs or ongoing runs are found, or the transfer was just created, trigger a manual run for the current time.
bq mk --transfer_run \
--transfer_config=<RESOURCE_NAME> \
--run_time=$(date -u +"%Y-%m-%dT%H:%M:%SZ")
Poll for Completion (5-Minute Rule): Attempt to check the status of the run every 30-60 seconds for up to 5 minutes.
bq ls --format=prettyjson --transfer_run --transfer_config=<RESOURCE_NAME>
Wait for User Guidance: Do NOT proceed until the user confirms ingestion is complete or provides guidance.
Once user confirms to proceed, start work on rest of the tasks.
deployment.yaml).