:mod:`airflow.contrib.operators.bigquery_to_bigquery` ===================================================== .. py:module:: airflow.contrib.operators.bigquery_to_bigquery Module Contents --------------- .. py:class:: BigQueryToBigQueryOperator(source_project_dataset_tables, destination_project_dataset_table, write_disposition='WRITE_EMPTY', create_disposition='CREATE_IF_NEEDED', bigquery_conn_id='bigquery_default', delegate_to=None, labels=None, *args, **kwargs) Bases: :class:`airflow.models.BaseOperator` Copies data from one BigQuery table to another. .. seealso:: For more details about these parameters: https://cloud.google.com/bigquery/docs/reference/v2/jobs#configuration.copy :param source_project_dataset_tables: One or more dotted ``(project:|project.).`` BigQuery tables to use as the source data. If ```` is not included, project will be the project defined in the connection json. Use a list if there are multiple source tables. (templated) :type source_project_dataset_tables: list|string :param destination_project_dataset_table: The destination BigQuery table. Format is: ``(project:|project.).
`` (templated) :type destination_project_dataset_table: str :param write_disposition: The write disposition if the table already exists. :type write_disposition: str :param create_disposition: The create disposition if the table doesn't exist. :type create_disposition: str :param bigquery_conn_id: reference to a specific BigQuery hook. :type bigquery_conn_id: str :param delegate_to: The account to impersonate, if any. For this to work, the service account making the request must have domain-wide delegation enabled. :type delegate_to: str :param labels: a dictionary containing labels for the job/query, passed to BigQuery :type labels: dict .. attribute:: template_fields :annotation: = ['source_project_dataset_tables', 'destination_project_dataset_table', 'labels'] .. attribute:: template_ext :annotation: = ['.sql'] .. attribute:: ui_color :annotation: = #e6f0e4 .. method:: execute(self, context)