Source code for

# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements.  See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership.  The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License.  You may obtain a copy of the License at
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# KIND, either express or implied.  See the License for the
# specific language governing permissions and limitations
# under the License.
"""This module contains a Google Cloud API base hook."""
from __future__ import annotations

import functools
import json
import logging
import os
import tempfile
import warnings
from contextlib import ExitStack, contextmanager
from subprocess import check_output
from typing import TYPE_CHECKING, Any, Callable, Generator, Sequence, TypeVar, cast

import google.auth
import google.auth.credentials
import google.oauth2.service_account
import google_auth_httplib2
import requests
import tenacity
from asgiref.sync import sync_to_async
from google.api_core.exceptions import Forbidden, ResourceExhausted, TooManyRequests
from google.auth import _cloud_sdk, compute_engine
from google.auth.environment_vars import CLOUD_SDK_CONFIG_DIR, CREDENTIALS
from google.auth.exceptions import RefreshError
from google.auth.transport import _http_client
from googleapiclient import discovery
from googleapiclient.errors import HttpError
from googleapiclient.http import MediaIoBaseDownload, build_http, set_user_agent

from airflow import version
from airflow.exceptions import AirflowException, AirflowProviderDeprecationWarning
from airflow.hooks.base import BaseHook
from import (
from import CLIENT_INFO
from airflow.utils.process_utils import patch_environ

    from google.api_core.gapic_v1.client_info import ClientInfo

[docs]log = logging.getLogger(__name__)
# Constants used by the mechanism of repeating requests in reaction to exceeding the temporary quota.
[docs]INVALID_KEYS = [ "DefaultRequestsPerMinutePerProject", "DefaultRequestsPerMinutePerUser", "RequestsPerMinutePerProject", "Resource has been exhausted (e.g. check quota).", ]
[docs]INVALID_REASONS = [ "userRateLimitExceeded", ]
[docs]def is_soft_quota_exception(exception: Exception): """ API for Google services does not have a standardized way to report quota violation errors. The function has been adapted by trial and error to the following services: * Google Translate * Google Vision * Google Text-to-Speech * Google Speech-to-Text * Google Natural Language * Google Video Intelligence """ if isinstance(exception, Forbidden): return any(reason in error.details() for reason in INVALID_REASONS for error in exception.errors) if isinstance(exception, (ResourceExhausted, TooManyRequests)): return any(key in error.details() for key in INVALID_KEYS for error in exception.errors) return False
[docs]def is_operation_in_progress_exception(exception: Exception) -> bool: """ Some calls return 429 (too many requests!) or 409 errors (Conflict) in case of operation in progress. * Google Cloud SQL """ if isinstance(exception, HttpError): return exception.resp.status == 429 or exception.resp.status == 409 return False
[docs]class retry_if_temporary_quota(tenacity.retry_if_exception): """Retries if there was an exception for exceeding the temporary quote limit.""" def __init__(self): super().__init__(is_soft_quota_exception)
[docs]class retry_if_operation_in_progress(tenacity.retry_if_exception): """Retries if there was an exception for exceeding the temporary quote limit.""" def __init__(self): super().__init__(is_operation_in_progress_exception)
# A fake project_id to use in functions decorated by fallback_to_default_project_id # This allows the 'project_id' argument to be of type str instead of str | None, # making it easier to type hint the function body without dealing with the None # case that can never happen at runtime.
[docs]PROVIDE_PROJECT_ID: str = cast(str, None)
[docs]T = TypeVar("T", bound=Callable)
[docs]RT = TypeVar("RT")
[docs]def get_field(extras: dict, field_name: str): """Get field from extra, first checking short name, then for backcompat we check for prefixed name.""" if field_name.startswith("extra__"): raise ValueError( f"Got prefixed name {field_name}; please remove the 'extra__google_cloud_platform__' prefix " "when using this method." ) if field_name in extras: return extras[field_name] or None prefixed_name = f"extra__google_cloud_platform__{field_name}" return extras.get(prefixed_name) or None
[docs]class GoogleBaseHook(BaseHook): """ A base hook for Google cloud-related hooks. Google cloud has a shared REST API client that is built in the same way no matter which service you use. This class helps construct and authorize the credentials needed to then call to actually discover and build a client for a Google cloud service. The class also contains some miscellaneous helper functions. All hook derived from this base hook use the 'Google Cloud' connection type. Three ways of authentication are supported: Default credentials: Only the 'Project Id' is required. You'll need to have set up default credentials, such as by the ``GOOGLE_APPLICATION_DEFAULT`` environment variable or from the metadata server on Google Compute Engine. JSON key file: Specify 'Project Id', 'Keyfile Path' and 'Scope'. Legacy P12 key files are not supported. JSON data provided in the UI: Specify 'Keyfile JSON'. :param gcp_conn_id: The connection ID to use when fetching connection info. :param delegate_to: The account to impersonate using domain-wide delegation of authority, if any. For this to work, the service account making the request must have domain-wide delegation enabled. The usage of this parameter should be limited only to Google Workspace (gsuite) and marketing platform operators and hooks. It is deprecated for usage by Google Cloud and Firebase operators and hooks, as well as transfer operators in other providers that involve Google cloud. :param impersonation_chain: Optional service account to impersonate using short-term credentials, or chained list of accounts required to get the access_token of the last account in the list, which will be impersonated in the request. If set as a string, the account must grant the originating account the Service Account Token Creator IAM role. If set as a sequence, the identities from the list must grant Service Account Token Creator IAM role to the directly preceding identity, with first account from the list granting this role to the originating account. """
[docs] conn_name_attr = "gcp_conn_id"
[docs] default_conn_name = "google_cloud_default"
[docs] conn_type = "google_cloud_platform"
[docs] hook_name = "Google Cloud"
[docs] def get_connection_form_widgets() -> dict[str, Any]: """Returns connection widgets to add to connection form.""" from flask_appbuilder.fieldwidgets import BS3PasswordFieldWidget, BS3TextFieldWidget from flask_babel import lazy_gettext from wtforms import IntegerField, PasswordField, StringField from wtforms.validators import NumberRange return { "project": StringField(lazy_gettext("Project Id"), widget=BS3TextFieldWidget()), "key_path": StringField(lazy_gettext("Keyfile Path"), widget=BS3TextFieldWidget()), "keyfile_dict": PasswordField(lazy_gettext("Keyfile JSON"), widget=BS3PasswordFieldWidget()), "credential_config_file": StringField( lazy_gettext("Credential Configuration File"), widget=BS3TextFieldWidget() ), "scope": StringField(lazy_gettext("Scopes (comma separated)"), widget=BS3TextFieldWidget()), "key_secret_name": StringField( lazy_gettext("Keyfile Secret Name (in GCP Secret Manager)"), widget=BS3TextFieldWidget() ), "key_secret_project_id": StringField( lazy_gettext("Keyfile Secret Project Id (in GCP Secret Manager)"), widget=BS3TextFieldWidget() ), "num_retries": IntegerField( lazy_gettext("Number of Retries"), validators=[NumberRange(min=0)], widget=BS3TextFieldWidget(), default=5, ), "impersonation_chain": StringField( lazy_gettext("Impersonation Chain"), widget=BS3TextFieldWidget() ), }
[docs] def get_ui_field_behaviour() -> dict[str, Any]: """Returns custom field behaviour.""" return { "hidden_fields": ["host", "schema", "login", "password", "port", "extra"], "relabeling": {}, }
def __init__( self, gcp_conn_id: str = "google_cloud_default", delegate_to: str | None = None, impersonation_chain: str | Sequence[str] | None = None, ) -> None: super().__init__() self.gcp_conn_id = gcp_conn_id self.delegate_to = delegate_to self.impersonation_chain = impersonation_chain self.extras: dict = self.get_connection(self.gcp_conn_id).extra_dejson self._cached_credentials: google.auth.credentials.Credentials | None = None self._cached_project_id: str | None = None
[docs] def get_credentials_and_project_id(self) -> tuple[google.auth.credentials.Credentials, str | None]: """Returns the Credentials object for Google API and the associated project_id.""" if self._cached_credentials is not None: return self._cached_credentials, self._cached_project_id key_path: str | None = self._get_field("key_path", None) try: keyfile_dict: str | dict[str, str] | None = self._get_field("keyfile_dict", None) keyfile_dict_json: dict[str, str] | None = None if keyfile_dict: if isinstance(keyfile_dict, dict): keyfile_dict_json = keyfile_dict else: keyfile_dict_json = json.loads(keyfile_dict) except json.decoder.JSONDecodeError: raise AirflowException("Invalid key JSON.") key_secret_name: str | None = self._get_field("key_secret_name", None) key_secret_project_id: str | None = self._get_field("key_secret_project_id", None) credential_config_file: str | None = self._get_field("credential_config_file", None) if not self.impersonation_chain: self.impersonation_chain = self._get_field("impersonation_chain", None) target_principal, delegates = _get_target_principal_and_delegates(self.impersonation_chain) credentials, project_id = get_credentials_and_project_id( key_path=key_path, keyfile_dict=keyfile_dict_json, credential_config_file=credential_config_file, key_secret_name=key_secret_name, key_secret_project_id=key_secret_project_id, scopes=self.scopes, delegate_to=self.delegate_to, target_principal=target_principal, delegates=delegates, ) overridden_project_id = self._get_field("project") if overridden_project_id: project_id = overridden_project_id self._cached_credentials = credentials self._cached_project_id = project_id return credentials, project_id
[docs] def get_credentials(self) -> google.auth.credentials.Credentials: """Returns the Credentials object for Google API.""" credentials, _ = self.get_credentials_and_project_id() return credentials
def _get_access_token(self) -> str: """Returns a valid access token from Google API Credentials.""" credentials = self.get_credentials() auth_req = google.auth.transport.requests.Request() # credentials.token is None # Need to refresh credentials to populate the token credentials.refresh(auth_req) return credentials.token @functools.lru_cache(maxsize=None) def _get_credentials_email(self) -> str: """ Returns the email address associated with the currently logged in account. If a service account is used, it returns the service account. If user authentication (e.g. gcloud auth) is used, it returns the e-mail account of that user. """ credentials = self.get_credentials() if isinstance(credentials, compute_engine.Credentials): try: credentials.refresh(_http_client.Request()) except RefreshError as msg: """ If the Compute Engine metadata service can't be reached in this case the instance has not credentials. """ self.log.debug(msg) service_account_email = getattr(credentials, "service_account_email", None) if service_account_email: return service_account_email http_authorized = self._authorize() oauth2_client ="oauth2", "v1", http=http_authorized, cache_discovery=False) return oauth2_client.tokeninfo().execute()["email"] def _authorize(self) -> google_auth_httplib2.AuthorizedHttp: """Returns an authorized HTTP object to be used to build a Google cloud service hook connection.""" credentials = self.get_credentials() http = build_http() http = set_user_agent(http, "airflow/" + version.version) authed_http = google_auth_httplib2.AuthorizedHttp(credentials, http=http) return authed_http def _get_field(self, f: str, default: Any = None) -> Any: """ Fetches a field from extras, and returns it. This is some Airflow magic. The google_cloud_platform hook type adds custom UI elements to the hook page, which allow admins to specify service_account, key_path, etc. They get formatted as shown below. """ return hasattr(self, "extras") and get_field(self.extras, f) or default @property
[docs] def project_id(self) -> str | None: """ Returns project id. :return: id of the project """ _, project_id = self.get_credentials_and_project_id() return project_id
[docs] def num_retries(self) -> int: """ Returns num_retries from Connection. :return: the number of times each API request should be retried """ field_value = self._get_field("num_retries", default=5) if field_value is None: return 5 if isinstance(field_value, str) and field_value.strip() == "": return 5 try: return int(field_value) except ValueError: raise AirflowException( f"The num_retries field should be a integer. " f'Current value: "{field_value}" (type: {type(field_value)}). ' f"Please check the connection configuration." )
[docs] def client_info(self) -> ClientInfo: """ Return client information used to generate a user-agent for API calls. It allows for better errors tracking. This object is only used by the google-cloud-* libraries that are built specifically for the Google Cloud. It is not supported by The Google APIs Python Client that use Discovery based APIs. """ warnings.warn( "This method is deprecated, please use ``.", AirflowProviderDeprecationWarning, stacklevel=2, ) return CLIENT_INFO
[docs] def scopes(self) -> Sequence[str]: """ Return OAuth 2.0 scopes. :return: Returns the scope defined in the connection configuration, or the default scope """ scope_value: str | None = self._get_field("scope", None) return _get_scopes(scope_value)
[docs] def quota_retry(*args, **kwargs) -> Callable: """Provides a mechanism to repeat requests in response to exceeding a temporary quota limit.""" def decorator(fun: Callable): default_kwargs = { "wait": tenacity.wait_exponential(multiplier=1, max=100), "retry": retry_if_temporary_quota(), "before": tenacity.before_log(log, logging.DEBUG), "after": tenacity.after_log(log, logging.DEBUG), } default_kwargs.update(**kwargs) return tenacity.retry(*args, **default_kwargs)(fun) return decorator
[docs] def operation_in_progress_retry(*args, **kwargs) -> Callable[[T], T]: """Provides a mechanism to repeat requests in response to operation in progress (HTTP 409) limit.""" def decorator(fun: T): default_kwargs = { "wait": tenacity.wait_exponential(multiplier=1, max=300), "retry": retry_if_operation_in_progress(), "before": tenacity.before_log(log, logging.DEBUG), "after": tenacity.after_log(log, logging.DEBUG), } default_kwargs.update(**kwargs) return cast(T, tenacity.retry(*args, **default_kwargs)(fun)) return decorator
[docs] def fallback_to_default_project_id(func: Callable[..., RT]) -> Callable[..., RT]: """ Decorator that provides fallback for Google Cloud project id. If the project is None it will be replaced with the project_id from the service account the Hook is authenticated with. Project id can be specified either via project_id kwarg or via first parameter in positional args. :param func: function to wrap :return: result of the function call """ @functools.wraps(func) def inner_wrapper(self: GoogleBaseHook, *args, **kwargs) -> RT: if args: raise AirflowException( "You must use keyword arguments in this methods rather than positional" ) if "project_id" in kwargs: kwargs["project_id"] = kwargs["project_id"] or self.project_id else: kwargs["project_id"] = self.project_id if not kwargs["project_id"]: raise AirflowException( "The project id must be passed either as " "keyword project_id parameter or as project_id extra " "in Google Cloud connection definition. Both are not set!" ) return func(self, *args, **kwargs) return inner_wrapper
[docs] def provide_gcp_credential_file(func: T) -> T: """ Provides a Google Cloud credentials for Application Default Credentials (ADC) strategy support. It is recommended to use ``provide_gcp_credential_file_as_context`` context manager to limit the scope when authorization data is available. Using context manager also makes it easier to use multiple connection in one function. """ @functools.wraps(func) def wrapper(self: GoogleBaseHook, *args, **kwargs): with self.provide_gcp_credential_file_as_context(): return func(self, *args, **kwargs) return cast(T, wrapper)
[docs] def provide_gcp_credential_file_as_context(self) -> Generator[str | None, None, None]: """ Provides a Google Cloud credentials for Application Default Credentials (ADC) strategy support. See: `Application Default Credentials (ADC) strategy <>`__. It can be used to provide credentials for external programs (e.g. gcloud) that expect authorization file in ``GOOGLE_APPLICATION_CREDENTIALS`` environment variable. """ key_path: str | None = self._get_field("key_path", None) keyfile_dict: str | dict[str, str] | None = self._get_field("keyfile_dict", None) if key_path and keyfile_dict: raise AirflowException( "The `keyfile_dict` and `key_path` fields are mutually exclusive. " "Please provide only one value." ) elif key_path: if key_path.endswith(".p12"): raise AirflowException("Legacy P12 key file are not supported, use a JSON key file.") with patch_environ({CREDENTIALS: key_path}): yield key_path elif keyfile_dict: with tempfile.NamedTemporaryFile(mode="w+t") as conf_file: if isinstance(keyfile_dict, dict): keyfile_dict = json.dumps(keyfile_dict) conf_file.write(keyfile_dict) conf_file.flush() with patch_environ({CREDENTIALS:}): yield else: # We will use the default service account credentials. yield None
[docs] def provide_authorized_gcloud(self) -> Generator[None, None, None]: """ Provides a separate gcloud configuration with current credentials. The gcloud tool allows you to login to Google Cloud only - ``gcloud auth login`` and for the needs of Application Default Credentials ``gcloud auth application-default login``. In our case, we want all commands to use only the credentials from ADCm so we need to configure the credentials in gcloud manually. """ credentials_path = _cloud_sdk.get_application_default_credentials_path() project_id = self.project_id with ExitStack() as exit_stack: exit_stack.enter_context(self.provide_gcp_credential_file_as_context()) gcloud_config_tmp = exit_stack.enter_context(tempfile.TemporaryDirectory()) exit_stack.enter_context(patch_environ({CLOUD_SDK_CONFIG_DIR: gcloud_config_tmp})) if CREDENTIALS in os.environ: # This solves most cases when we are logged in using the service key in Airflow. # Don't display stdout/stderr for security reason check_output( [ "gcloud", "auth", "activate-service-account", f"--key-file={os.environ[CREDENTIALS]}", ] ) elif os.path.exists(credentials_path): # If we are logged in by `gcloud auth application-default` then we need to log in manually. # This will make the `gcloud auth application-default` and `gcloud auth` credentials equals. with open(credentials_path) as creds_file: creds_content = json.loads( # Don't display stdout/stderr for security reason check_output(["gcloud", "config", "set", "auth/client_id", creds_content["client_id"]]) # Don't display stdout/stderr for security reason check_output( ["gcloud", "config", "set", "auth/client_secret", creds_content["client_secret"]] ) # Don't display stdout/stderr for security reason check_output( [ "gcloud", "auth", "activate-refresh-token", creds_content["client_id"], creds_content["refresh_token"], ] ) if project_id: # Don't display stdout/stderr for security reason check_output(["gcloud", "config", "set", "core/project", project_id]) yield
[docs] def download_content_from_request(file_handle, request: dict, chunk_size: int) -> None: """ Download media resources. Note that the Python file object is compatible with io.Base and can be used with this class also. :param file_handle: io.Base or file object. The stream in which to write the downloaded bytes. :param request: googleapiclient.http.HttpRequest, the media request to perform in chunks. :param chunk_size: int, File will be downloaded in chunks of this many bytes. """ downloader = MediaIoBaseDownload(file_handle, request, chunksize=chunk_size) done = False while done is False: _, done = downloader.next_chunk() file_handle.flush()
[docs] def test_connection(self): """Test the Google cloud connectivity from UI.""" status, message = False, "" try: token = self._get_access_token() url = f"{token}" response = if response.status_code == 200: status = True message = "Connection successfully tested" except Exception as e: status = False message = str(e) return status, message
[docs]class GoogleBaseAsyncHook(BaseHook): """GoogleBaseAsyncHook inherits from BaseHook class, run on the trigger worker."""
[docs] sync_hook_class: Any = None
def __init__(self, **kwargs: Any): self._hook_kwargs = kwargs self._sync_hook = None
[docs] async def get_sync_hook(self) -> Any: """Sync version of the Google Cloud Hook makes blocking calls in ``__init__``; don't inherit it.""" if not self._sync_hook: self._sync_hook = await sync_to_async(self.sync_hook_class)(**self._hook_kwargs) return self._sync_hook
[docs] async def service_file_as_context(self) -> Any: sync_hook = await self.get_sync_hook() return await sync_to_async(sync_hook.provide_gcp_credential_file_as_context)()

Was this entry helpful?