:mod:`airflow.operators.s3_to_redshift_operator` ================================================ .. py:module:: airflow.operators.s3_to_redshift_operator Module Contents --------------- .. py:class:: S3ToRedshiftTransfer(schema, table, s3_bucket, s3_key, redshift_conn_id='redshift_default', aws_conn_id='aws_default', verify=None, copy_options=tuple(), autocommit=False, parameters=None, *args, **kwargs) Bases: :class:`airflow.models.BaseOperator` Executes an COPY command to load files from s3 to Redshift :param schema: reference to a specific schema in redshift database :type schema: str :param table: reference to a specific table in redshift database :type table: str :param s3_bucket: reference to a specific S3 bucket :type s3_bucket: str :param s3_key: reference to a specific S3 key :type s3_key: str :param redshift_conn_id: reference to a specific redshift database :type redshift_conn_id: str :param aws_conn_id: reference to a specific S3 connection :type aws_conn_id: str :param verify: Whether or not to verify SSL certificates for S3 connection. By default SSL certificates are verified. You can provide the following values: - ``False``: do not validate SSL certificates. SSL will still be used (unless use_ssl is False), but SSL certificates will not be verified. - ``path/to/cert/bundle.pem``: A filename of the CA cert bundle to uses. You can specify this argument if you want to use a different CA cert bundle than the one used by botocore. :type verify: bool or str :param copy_options: reference to a list of COPY options :type copy_options: list .. attribute:: template_fields :annotation: = [] .. attribute:: template_ext :annotation: = [] .. attribute:: ui_color :annotation: = #ededed .. method:: execute(self, context)