def bigquery_dq_check()

in utils/gcp.py [0:0]


def bigquery_dq_check(
    source_table,
    dataset_id,
    task_id,
    parameters=(),
    arguments=(),
    project_id="moz-fx-data-shared-prod",
    gcp_conn_id="google_cloud_airflow_gke",
    gke_project_id=GCP_PROJECT_ID,
    gke_location="us-west1",
    gke_cluster_name="workloads-prod-v1",
    gke_namespace="default",
    docker_image=BIGQUERY_ETL_DOCKER_IMAGE,
    date_partition_parameter="submission_date",
    is_dq_check_fail=True,
    **kwargs,