Source code for tests.system.providers.amazon.aws.example_rds_export
# Licensed to the Apache Software Foundation (ASF) under one# or more contributor license agreements. See the NOTICE file# distributed with this work for additional information# regarding copyright ownership. The ASF licenses this file# to you under the Apache License, Version 2.0 (the# "License"); you may not use this file except in compliance# with the License. You may obtain a copy of the License at## http://www.apache.org/licenses/LICENSE-2.0## Unless required by applicable law or agreed to in writing,# software distributed under the License is distributed on an# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY# KIND, either express or implied. See the License for the# specific language governing permissions and limitations# under the License.from__future__importannotationsfromdatetimeimportdatetimefromairflowimportDAGfromairflow.decoratorsimporttaskfromairflow.models.baseoperatorimportchainfromairflow.providers.amazon.aws.hooks.rdsimportRdsHookfromairflow.providers.amazon.aws.operators.rdsimport(RdsCancelExportTaskOperator,RdsCreateDbInstanceOperator,RdsCreateDbSnapshotOperator,RdsDeleteDbInstanceOperator,RdsDeleteDbSnapshotOperator,RdsStartExportTaskOperator,)fromairflow.providers.amazon.aws.operators.s3importS3CreateBucketOperator,S3DeleteBucketOperatorfromairflow.providers.amazon.aws.sensors.rdsimportRdsExportTaskExistenceSensor,RdsSnapshotExistenceSensorfromairflow.utils.trigger_ruleimportTriggerRulefromtests.system.providers.amazon.aws.utilsimportENV_ID_KEY,SystemTestContextBuilder
env_id=test_context[ENV_ID_KEY]bucket_name:str=f"{env_id}-bucket"rds_db_name:str=f"{env_id}_db"rds_instance_name:str=f"{env_id}-instance"rds_snapshot_name:str=f"{env_id}-snapshot"rds_export_task_id:str=f"{env_id}-export-task"create_bucket=S3CreateBucketOperator(task_id="create_bucket",bucket_name=bucket_name,)create_db_instance=RdsCreateDbInstanceOperator(task_id="create_db_instance",db_instance_identifier=rds_instance_name,db_instance_class="db.t4g.micro",engine="postgres",rds_kwargs={"MasterUsername":"rds_username",# NEVER store your production password in plaintext in a DAG like this.# Use Airflow Secrets or a secret manager for this in production."MasterUserPassword":"rds_password","AllocatedStorage":20,"DBName":rds_db_name,"PubliclyAccessible":False,},)create_snapshot=RdsCreateDbSnapshotOperator(task_id="create_snapshot",db_type="instance",db_identifier=rds_instance_name,db_snapshot_identifier=rds_snapshot_name,)await_snapshot=RdsSnapshotExistenceSensor(task_id="snapshot_sensor",db_type="instance",db_snapshot_identifier=rds_snapshot_name,target_statuses=["available"],)snapshot_arn=get_snapshot_arn(rds_snapshot_name)# [START howto_operator_rds_start_export_task]start_export=RdsStartExportTaskOperator(task_id="start_export",export_task_identifier=rds_export_task_id,source_arn=snapshot_arn,s3_bucket_name=bucket_name,s3_prefix="rds-test",iam_role_arn=test_context[ROLE_ARN_KEY],kms_key_id=test_context[KMS_KEY_ID_KEY],)# [END howto_operator_rds_start_export_task]# RdsStartExportTaskOperator waits by default, setting as False to test the Sensor below.start_export.wait_for_completion=False# [START howto_operator_rds_cancel_export]cancel_export=RdsCancelExportTaskOperator(task_id="cancel_export",export_task_identifier=rds_export_task_id,)# [END howto_operator_rds_cancel_export]cancel_export.check_interval=10# [START howto_sensor_rds_export_task_existence]export_sensor=RdsExportTaskExistenceSensor(task_id="export_sensor",export_task_identifier=rds_export_task_id,target_statuses=["canceled"],)# [END howto_sensor_rds_export_task_existence]delete_snapshot=RdsDeleteDbSnapshotOperator(task_id="delete_snapshot",db_type="instance",db_snapshot_identifier=rds_snapshot_name,trigger_rule=TriggerRule.ALL_DONE,)delete_bucket=S3DeleteBucketOperator(task_id="delete_bucket",bucket_name=bucket_name,force_delete=True,trigger_rule=TriggerRule.ALL_DONE,)delete_db_instance=RdsDeleteDbInstanceOperator(task_id="delete_db_instance",db_instance_identifier=rds_instance_name,rds_kwargs={"SkipFinalSnapshot":True},trigger_rule=TriggerRule.ALL_DONE,)chain(# TEST SETUPtest_context,create_bucket,create_db_instance,create_snapshot,await_snapshot,snapshot_arn,# TEST BODYstart_export,cancel_export,export_sensor,# TEST TEARDOWNdelete_snapshot,delete_bucket,delete_db_instance,)fromtests.system.utils.watcherimportwatcher# This test needs watcher in order to properly mark success/failure# when "tearDown" task with trigger rule is part of the DAGlist(dag.tasks)>>watcher()fromtests.system.utilsimportget_test_run# noqa: E402# Needed to run the example DAG with pytest (see: tests/system/README.md#run_via_pytest)