Source code for airflow.contrib.hooks.datastore_hook

# -*- coding: utf-8 -*-
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# See the License for the specific language governing permissions and
# limitations under the License.

import json
import time
from apiclient.discovery import build
from airflow.contrib.hooks.gcp_api_base_hook import GoogleCloudBaseHook

[docs]class DatastoreHook(GoogleCloudBaseHook): """ Interact with Google Cloud Datastore. This hook uses the Google Cloud Platform connection. This object is not threads safe. If you want to make multiple requests simultaniously, you will need to create a hook per thread. """ def __init__(self, datastore_conn_id='google_cloud_datastore_default', delegate_to=None): super(DatastoreHook, self).__init__(datastore_conn_id, delegate_to) self.connection = self.get_conn() self.admin_connection = self.get_conn('v1beta1')
[docs] def get_conn(self, version='v1'): """ Returns a Google Cloud Storage service object. """ http_authorized = self._authorize() return build('datastore', version, http=http_authorized)
[docs] def allocate_ids(self, partialKeys): """ Allocate IDs for incomplete keys. see :param partialKeys: a list of partial keys :return: a list of full keys. """ resp = self.connection.projects().allocateIds(projectId=self.project_id, body={'keys': partialKeys}).execute() return resp['keys']
[docs] def begin_transaction(self): """ Get a new transaction handle see :return: a transaction handle """ resp = self.connection.projects().beginTransaction(projectId=self.project_id, body={}).execute() return resp['transaction']
[docs] def commit(self, body): """ Commit a transaction, optionally creating, deleting or modifying some entities. see :param body: the body of the commit request :return: the response body of the commit request """ resp = self.connection.projects().commit(projectId=self.project_id, body=body).execute() return resp
[docs] def lookup(self, keys, read_consistency=None, transaction=None): """ Lookup some entities by key see :param keys: the keys to lookup :param read_consistency: the read consistency to use. default, strong or eventual. Cannot be used with a transaction. :param transaction: the transaction to use, if any. :return: the response body of the lookup request. """ body = {'keys': keys} if read_consistency: body['readConsistency'] = read_consistency if transaction: body['transaction'] = transaction return self.connection.projects().lookup(projectId=self.project_id, body=body).execute()
[docs] def rollback(self, transaction): """ Roll back a transaction see :param transaction: the transaction to roll back """ self.connection.projects().rollback(projectId=self.project_id, body={'transaction': transaction})\ .execute()
[docs] def run_query(self, body): """ Run a query for entities. see :param body: the body of the query request :return: the batch of query results. """ resp = self.connection.projects().runQuery(projectId=self.project_id, body=body).execute() return resp['batch']
[docs] def get_operation(self, name): """ Gets the latest state of a long-running operation :param name: the name of the operation resource """ resp = self.connection.projects().operations().get(name=name).execute() return resp
[docs] def delete_operation(self, name): """ Deletes the long-running operation :param name: the name of the operation resource """ resp = self.connection.projects().operations().delete(name=name).execute() return resp
[docs] def poll_operation_until_done(self, name, polling_interval_in_seconds): """ Poll backup operation state until it's completed """ while True: result = self.get_operation(name) state = result['metadata']['common']['state'] if state == 'PROCESSING':'Operation is processing. Re-polling state in {} seconds' .format(polling_interval_in_seconds)) time.sleep(polling_interval_in_seconds) else: return result
[docs] def export_to_storage_bucket(self, bucket, namespace=None, entity_filter=None, labels=None): """ Export entities from Cloud Datastore to Cloud Storage for backup """ output_uri_prefix = 'gs://' + ('/').join(filter(None, [bucket, namespace])) if not entity_filter: entity_filter = {} if not labels: labels = {} body = { 'outputUrlPrefix': output_uri_prefix, 'entityFilter': entity_filter, 'labels': labels, } resp = self.admin_connection.projects().export(projectId=self.project_id, body=body).execute() return resp
[docs] def import_from_storage_bucket(self, bucket, file, namespace=None, entity_filter=None, labels=None): """ Import a backup from Cloud Storage to Cloud Datastore """ input_url = 'gs://' + ('/').join(filter(None, [bucket, namespace, file])) if not entity_filter: entity_filter = {} if not labels: labels = {} body = { 'inputUrl': input_url, 'entityFilter': entity_filter, 'labels': labels, } resp = self.admin_connection.projects().import_(projectId=self.project_id, body=body).execute() return resp