gcp-dataflow
Installation
SKILL.md
Apache Beam Pipelines on Cloud Dataflow
Expert guidance for writing and packaging Apache Beam pipelines to run on Google Cloud Dataflow.
Creating a new project
Use this section when creating a new project for a Dataflow pipeline.
- If the user doesn't say explicitly which language (Java, Python, Go) shall be used to write the pipeline, you MUST confirm the language.
- Determine which version of Beam SDK should be used by searching for the most
recently released version of Apache Beam, unless the user already uses a
particular version.
- Action: Run a web search for the latest Apache Beam SDK release.
- YOU MUST use same version of Apache Beam consistently throughout the project
in Dockerfiles,
requirements.txt, and other similar files where versions are specified.
Related skills
More from gemini-cli-extensions/data-agent-kit-starter-pack
gcp-spark
|
7dbt-bigquery
Expert guidance for creating, modifying, and optimizing dbt pipelines
7dataform-bigquery
Expertise in generating clean, correct, and efficient Dataform pipeline
7data-autocleaning
Automated data quality and transformation capabilities for Dataform/dbt/BigQuery
7ml-best-practices
|
7building-data-apps
|
7