airflow.providers.google.cloud.transfers.bigquery_to_gcs¶
This module contains Google BigQuery to Google Cloud Storage operator.
Module Contents¶
Classes¶
| Transfers a BigQuery table to a Google Cloud Storage bucket. | 
- class airflow.providers.google.cloud.transfers.bigquery_to_gcs.BigQueryToGCSOperator(*, source_project_dataset_table, destination_cloud_storage_uris, project_id=None, compression='NONE', export_format='CSV', field_delimiter=',', print_header=True, gcp_conn_id='google_cloud_default', labels=None, location=None, impersonation_chain=None, result_retry=DEFAULT_RETRY, result_timeout=None, job_id=None, force_rerun=False, reattach_states=None, deferrable=False, **kwargs)[source]¶
- Bases: - airflow.models.BaseOperator- Transfers a BigQuery table to a Google Cloud Storage bucket. - See also - For more information on how to use this operator, take a look at the guide: Operator - See also - For more details about these parameters: https://cloud.google.com/bigquery/docs/reference/v2/jobs - Parameters
- source_project_dataset_table (str) – The dotted - (<project>.|<project>:)<dataset>.<table>BigQuery table to use as the source data. If- <project>is not included, project will be the project defined in the connection json. (templated)
- destination_cloud_storage_uris (list[str]) – The destination Google Cloud Storage URI (e.g. gs://some-bucket/some-file.txt). (templated) Follows convention defined here: https://cloud.google.com/bigquery/exporting-data-from-bigquery#exportingmultiple 
- project_id (str | None) – Google Cloud Project where the job is running 
- compression (str) – Type of compression to use. 
- export_format (str) – File format to export. 
- field_delimiter (str) – The delimiter to use when extracting to a CSV. 
- print_header (bool) – Whether to print a header for a CSV file extract. 
- gcp_conn_id (str) – (Optional) The connection ID used to connect to Google Cloud. 
- labels (dict | None) – a dictionary containing labels for the job/query, passed to BigQuery 
- location (str | None) – The location used for the operation. 
- impersonation_chain (str | Sequence[str] | None) – Optional service account to impersonate using short-term credentials, or chained list of accounts required to get the access_token of the last account in the list, which will be impersonated in the request. If set as a string, the account must grant the originating account the Service Account Token Creator IAM role. If set as a sequence, the identities from the list must grant Service Account Token Creator IAM role to the directly preceding identity, with first account from the list granting this role to the originating account (templated). 
- result_retry (google.api_core.retry.Retry) – How to retry the result call that retrieves rows 
- result_timeout (float | None) – The number of seconds to wait for result method before using result_retry 
- job_id (str | None) – The ID of the job. It will be suffixed with hash of job configuration unless - force_rerunis True. The ID must contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or dashes (-). The maximum length is 1,024 characters. If not provided then uuid will be generated.
- force_rerun (bool) – If True then operator will use hash of uuid as job id suffix 
- reattach_states (set[str] | None) – Set of BigQuery job’s states in case of which we should reattach to the job. Should be other than final states. 
- deferrable (bool) – Run operator in the deferrable mode 
 
 - template_fields: Sequence[str] = ('source_project_dataset_table', 'destination_cloud_storage_uris', 'export_format', 'labels',...[source]¶
 
