This is an automated email from the ASF dual-hosted git repository.
potiuk pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/airflow.git
The following commit(s) were added to refs/heads/main by this push:
new 6ab02b69a0 Remove bigquery example already migrated to AIP-47 (#24379)
6ab02b69a0 is described below
commit 6ab02b69a0bbc398e9e41c70d78487d3255b0585
Author: Chenglong Yan <[email protected]>
AuthorDate: Sun Jun 12 17:55:12 2022 +0800
Remove bigquery example already migrated to AIP-47 (#24379)
related: #22430, #22447, #22311
---
.../cloud/example_dags/example_bigquery_to_gcs.py | 68 ----------------------
.../cloud/transfers/test_bigquery_to_gcs_system.py | 42 -------------
2 files changed, 110 deletions(-)
diff --git
a/airflow/providers/google/cloud/example_dags/example_bigquery_to_gcs.py
b/airflow/providers/google/cloud/example_dags/example_bigquery_to_gcs.py
deleted file mode 100644
index ba66e10e44..0000000000
--- a/airflow/providers/google/cloud/example_dags/example_bigquery_to_gcs.py
+++ /dev/null
@@ -1,68 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-"""
-Example Airflow DAG for Google BigQuery service.
-"""
-import os
-from datetime import datetime
-
-from airflow import models
-from airflow.providers.google.cloud.operators.bigquery import (
- BigQueryCreateEmptyDatasetOperator,
- BigQueryCreateEmptyTableOperator,
- BigQueryDeleteDatasetOperator,
-)
-from airflow.providers.google.cloud.transfers.bigquery_to_gcs import
BigQueryToGCSOperator
-
-PROJECT_ID = os.environ.get("GCP_PROJECT_ID", "example-project")
-DATASET_NAME = os.environ.get("GCP_BIGQUERY_DATASET_NAME",
"test_dataset_transfer")
-DATA_EXPORT_BUCKET_NAME = os.environ.get("GCP_BIGQUERY_EXPORT_BUCKET_NAME",
"INVALID BUCKET NAME")
-TABLE = "table_42"
-
-with models.DAG(
- "example_bigquery_to_gcs",
- schedule_interval=None, # Override to match your needs
- start_date=datetime(2021, 1, 1),
- catchup=False,
- tags=["example"],
-) as dag:
- bigquery_to_gcs = BigQueryToGCSOperator(
- task_id="bigquery_to_gcs",
- source_project_dataset_table=f"{DATASET_NAME}.{TABLE}",
-
destination_cloud_storage_uris=[f"gs://{DATA_EXPORT_BUCKET_NAME}/export-bigquery.csv"],
- )
-
- create_dataset =
BigQueryCreateEmptyDatasetOperator(task_id="create_dataset",
dataset_id=DATASET_NAME)
-
- create_table = BigQueryCreateEmptyTableOperator(
- task_id="create_table",
- dataset_id=DATASET_NAME,
- table_id=TABLE,
- schema_fields=[
- {"name": "emp_name", "type": "STRING", "mode": "REQUIRED"},
- {"name": "salary", "type": "INTEGER", "mode": "NULLABLE"},
- ],
- )
- create_dataset >> create_table >> bigquery_to_gcs
-
- delete_dataset = BigQueryDeleteDatasetOperator(
- task_id="delete_dataset", dataset_id=DATASET_NAME, delete_contents=True
- )
-
- bigquery_to_gcs >> delete_dataset
diff --git
a/tests/providers/google/cloud/transfers/test_bigquery_to_gcs_system.py
b/tests/providers/google/cloud/transfers/test_bigquery_to_gcs_system.py
deleted file mode 100644
index 6887572c0f..0000000000
--- a/tests/providers/google/cloud/transfers/test_bigquery_to_gcs_system.py
+++ /dev/null
@@ -1,42 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements. See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership. The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License. You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied. See the License for the
-# specific language governing permissions and limitations
-# under the License.
-"""System tests for Google Cloud Build operators"""
-import pytest
-
-from airflow.providers.google.cloud.example_dags.example_bigquery_to_gcs
import DATA_EXPORT_BUCKET_NAME
-from tests.providers.google.cloud.utils.gcp_authenticator import
GCP_BIGQUERY_KEY
-from tests.test_utils.gcp_system_helpers import CLOUD_DAG_FOLDER,
GoogleSystemTest, provide_gcp_context
-
-
[email protected]("mysql", "postgres")
[email protected]("google.cloud")
[email protected]_file(GCP_BIGQUERY_KEY)
-class BigQueryToGCSExampleDagsSystemTest(GoogleSystemTest):
- @provide_gcp_context(GCP_BIGQUERY_KEY)
- def setUp(self):
- super().setUp()
- self.create_gcs_bucket(DATA_EXPORT_BUCKET_NAME)
-
- @provide_gcp_context(GCP_BIGQUERY_KEY)
- def test_run_example_dag_queries(self):
- self.run_dag('example_bigquery_to_gcs', CLOUD_DAG_FOLDER)
-
- @provide_gcp_context(GCP_BIGQUERY_KEY)
- def tearDown(self):
- self.delete_gcs_bucket(DATA_EXPORT_BUCKET_NAME)
- super().tearDown()