datasets/eumetsat/pipelines/solar_forecasting/pipeline.yaml (44 lines of code) (raw):
resources: ~
dag:
airflow_version: 2
initialize:
dag_id: solar_forecasting
default_args:
owner: "Google"
depends_on_past: False
start_date: '2022-01-19'
max_active_runs: 1
schedule_interval: "0 1 * * 6"
catchup: False
default_view: graph
tasks:
- operator: "CloudDataTransferServiceGCSToGCSOperator"
description: "Task to run a GCS to GCS operation using Google resources"
args:
task_id: copy_gcs_bucket_v3
timeout: 172800 # 24 hours
retries: 0
wait: True
project_id: bigquery-public-data
source_bucket: "{{ var.json.eumetsat.solar_forecasting.source_bucket }}"
object_conditions:
includePrefixes:
- satellite/EUMETSAT/SEVIRI_RSS/v3
destination_bucket: "{{ var.json.eumetsat.solar_forecasting.destination_bucket }}"
google_impersonation_chain: "{{ var.json.eumetsat.solar_forecasting.service_account }}"
- operator: "CloudDataTransferServiceGCSToGCSOperator"
description: "Task to run a GCS to GCS operation using Google resources"
args:
task_id: copy_gcs_bucket_v4
timeout: 172800 # 24 hours
retries: 0
wait: True
project_id: bigquery-public-data
source_bucket: "{{ var.json.eumetsat.solar_forecasting.source_bucket }}"
object_conditions:
includePrefixes:
- satellite/EUMETSAT/SEVIRI_RSS/v4
destination_bucket: "{{ var.json.eumetsat.solar_forecasting.destination_bucket }}"
google_impersonation_chain: "{{ var.json.eumetsat.solar_forecasting.service_account }}"
graph_paths:
- "copy_gcs_bucket_v3 >> copy_gcs_bucket_v4"