tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor

Example Airflow DAG that show how to check Hive partitions existence using Dataproc Metastore Sensor.

Note that Metastore service must be configured to use gRPC endpoints.

Module Contents

tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.DAG_ID = 'hive_partition_sensor'[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.PROJECT_ID[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.ENV_ID[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.REGION = 'us-central1'[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.NETWORK = 'default'[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.METASTORE_SERVICE_ID[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.METASTORE_TIMEOUT = 2400[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.METASTORE_SERVICE[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.METASTORE_SERVICE_QFN[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.DATAPROC_CLUSTER_NAME[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.DATAPROC_CLUSTER_CONFIG[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.TABLE_NAME = 'transactions_partitioned'[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.COLUMN = 'TransactionType'[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.PARTITION_1[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.PARTITION_2[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.SOURCE_DATA_BUCKET = 'airflow-system-tests-resources'[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.SOURCE_DATA_PATH = 'dataproc/hive'[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.SOURCE_DATA_FILE_NAME = 'part-00000.parquet'[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.EXTERNAL_TABLE_BUCKET = "{{task_instance.xcom_pull(task_ids='get_hive_warehouse_bucket_task', key='bucket')}}"[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.QUERY_CREATE_EXTERNAL_TABLE[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.QUERY_CREATE_PARTITIONED_TABLE[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.QUERY_COPY_DATA_WITH_PARTITIONS[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.create_metastore_service[source]
tests.system.providers.google.cloud.dataproc_metastore.example_dataproc_metastore_hive_partition_sensor.test_run[source]

Was this entry helpful?