From data-agent-kit-starter-pack
Guides writing, packaging, and executing Apache Beam pipelines on GCP Dataflow, including Flex Templates for Java, Python, and Go projects.
npx claudepluginhub gemini-cli-extensions/data-agent-kit-starter-pack --plugin data-agent-kit-starter-packThis skill uses the workspace's default tool permissions.
Expert guidance for writing and packaging Apache Beam pipelines to run on Google
Implements structured self-debugging workflow for AI agent failures: capture errors, diagnose patterns like loops or context overflow, apply contained recoveries, and generate introspection reports.
Monitors deployed URLs for regressions in HTTP status, console errors, performance metrics, content, network, and APIs after deploys, merges, or upgrades.
Provides React and Next.js patterns for component composition, compound components, state management, data fetching, performance optimization, forms, routing, and accessible UIs.
Expert guidance for writing and packaging Apache Beam pipelines to run on Google Cloud Dataflow.
Use this section when creating a new project for a Dataflow pipeline.
requirements.txt, and other similar files where versions
are specified.Use this section when configuring a Dataflow Java pipeline project using gradle.
com.github.johnrengelman.shadow) unless the user explicitly requests a
Fat Jar.application plugin for
passing command-line parameters.slf4j-api version pulled transitively by Apache Beam.slf4j-simple,
logback-classic, etc.) to exactly match the major/minor version of the
resolved slf4j-api.When creating new Dataflow pipeline projects, configure them as a Flex template.
Flex Templates offer a hermetic and reproducible launch environment, and are
easy to launch with gcloud or with orchestrators like Cloud Composer.
Follow the Flex Templates section below.
--sdk_container_image). Whenever
configuring or suggesting a Dataflow Flex Template for a Python pipeline
that requires extra dependencies (e.g., using --requirements_file,
--setup_file, or --extra_package), YOU MUST recommend the Single
Docker Image Configuration as detailed in
python_flex_template_reference.md.cloudbuild.yaml out-of-the-box for
building and pushing images unless local setup is explicitly requested.When launching Python Pipelines without a Flex Template with
DataflowRunner, you MUST scan the pipeline project directory for the
following files:
requirements.txt:
--requirements_file pipeline option.setup.py:
--setup_file pipeline option. This is
critical if the pipeline uses local modules or packages.When launching Python Pipelines with a Flex Template, if the Flex Template
image is also the SDK Container image (Single Docker Image Configuration),
then you MUST supply the image in the sdk_container_image parameter.
Confirm the launch command with the user.
your-gcp-project-id) for GCP
resources when drafting run scripts or configs. Action: If values are
unknown, proactively run commands like gcloud config get-value project to
find active resources to pre-fill scripts for the user. Confirm the values
with the user before proceeding.