airflow.operators.s3_to_redshift_operator¶
Module Contents¶
- 
class airflow.operators.s3_to_redshift_operator.S3ToRedshiftTransfer(schema, table, s3_bucket, s3_key, redshift_conn_id='redshift_default', aws_conn_id='aws_default', verify=None, copy_options=tuple(), autocommit=False, parameters=None, *args, **kwargs)[source]¶
- Bases: - airflow.models.BaseOperator- Executes an COPY command to load files from s3 to Redshift - Parameters
- schema (str) – reference to a specific schema in redshift database 
- table (str) – reference to a specific table in redshift database 
- s3_bucket (str) – reference to a specific S3 bucket 
- s3_key (str) – reference to a specific S3 key 
- redshift_conn_id (str) – reference to a specific redshift database 
- aws_conn_id (str) – reference to a specific S3 connection 
- Whether or not to verify SSL certificates for S3 connection. By default SSL certificates are verified. You can provide the following values: - False: do not validate SSL certificates. SSL will still be used
- (unless use_ssl is False), but SSL certificates will not be verified. 
 
- path/to/cert/bundle.pem: A filename of the CA cert bundle to uses.
- You can specify this argument if you want to use a different CA cert bundle than the one used by botocore. 
 
 
- copy_options (list) – reference to a list of COPY options