airflow.providers.amazon.aws.operators.s3_bucket¶
This module contains AWS S3 operators.
Module Contents¶
- 
class airflow.providers.amazon.aws.operators.s3_bucket.S3CreateBucketOperator(*, bucket_name: str, aws_conn_id: Optional[str] = 'aws_default', region_name: Optional[str] = None, **kwargs)[source]¶
- Bases: - airflow.models.BaseOperator- This operator creates an S3 bucket - Parameters
- bucket_name (str) -- This is bucket name you want to create 
- aws_conn_id (Optional[str]) -- The Airflow connection used for AWS credentials. If this is None or empty then the default boto3 behaviour is used. If running Airflow in a distributed manner and aws_conn_id is None or empty, then default boto3 configuration would be used (and must be maintained on each worker node). 
- region_name (Optional[str]) -- AWS region_name. If not specified fetched from connection. 
 
 
- 
class airflow.providers.amazon.aws.operators.s3_bucket.S3DeleteBucketOperator(bucket_name: str, force_delete: bool = False, aws_conn_id: Optional[str] = 'aws_default', **kwargs)[source]¶
- Bases: - airflow.models.BaseOperator- This operator deletes an S3 bucket - Parameters
- bucket_name (str) -- This is bucket name you want to delete 
- force_delete (bool) -- Forcibly delete all objects in the bucket before deleting the bucket 
- aws_conn_id (Optional[str]) -- The Airflow connection used for AWS credentials. If this is None or empty then the default boto3 behaviour is used. If running Airflow in a distributed manner and aws_conn_id is None or empty, then default boto3 configuration would be used (and must be maintained on each worker node).