in utils/gcp.py [0:0]
def bigquery_dq_check(
source_table,
dataset_id,
task_id,
parameters=(),
arguments=(),
project_id="moz-fx-data-shared-prod",
gcp_conn_id="google_cloud_airflow_gke",
gke_project_id=GCP_PROJECT_ID,
gke_location="us-west1",
gke_cluster_name="workloads-prod-v1",
gke_namespace="default",
docker_image=BIGQUERY_ETL_DOCKER_IMAGE,
date_partition_parameter="submission_date",
is_dq_check_fail=True,
**kwargs,