:mod:`airflow.contrib.operators.datastore_import_operator` ========================================================== .. py:module:: airflow.contrib.operators.datastore_import_operator Module Contents --------------- .. py:class:: DatastoreImportOperator(bucket, file, namespace=None, entity_filter=None, labels=None, datastore_conn_id='google_cloud_default', delegate_to=None, polling_interval_in_seconds=10, xcom_push=False, *args, **kwargs) Bases: :class:`airflow.models.BaseOperator` Import entities from Cloud Storage to Google Cloud Datastore :param bucket: container in Cloud Storage to store data :type bucket: str :param file: path of the backup metadata file in the specified Cloud Storage bucket. It should have the extension .overall_export_metadata :type file: str :param namespace: optional namespace of the backup metadata file in the specified Cloud Storage bucket. :type namespace: str :param entity_filter: description of what data from the project is included in the export, refer to https://cloud.google.com/datastore/docs/reference/rest/Shared.Types/EntityFilter :type entity_filter: dict :param labels: client-assigned labels for cloud storage :type labels: dict :param datastore_conn_id: the name of the connection id to use :type datastore_conn_id: str :param delegate_to: The account to impersonate, if any. For this to work, the service account making the request must have domain-wide delegation enabled. :type delegate_to: str :param polling_interval_in_seconds: number of seconds to wait before polling for execution status again :type polling_interval_in_seconds: int :param xcom_push: push operation name to xcom for reference :type xcom_push: bool .. method:: execute(self, context)