airflow.providers.databricks.operators.databricks_repos

This module contains Databricks operators.

Module Contents

Classes

DatabricksReposUpdateOperator

Updates specified repository to a given branch or tag using

class airflow.providers.databricks.operators.databricks_repos.DatabricksReposUpdateOperator(*, branch=None, tag=None, repo_id=None, repo_path=None, databricks_conn_id='databricks_default', databricks_retry_limit=3, databricks_retry_delay=1, **kwargs)[source]

Bases: airflow.models.BaseOperator

Updates specified repository to a given branch or tag using api/2.0/repos/ API endpoint.

Parameters
  • branch (Optional[str]) -- optional name of branch to update to. Should be specified if tag is omitted

  • tag (Optional[str]) -- optional name of tag to update to. Should be specified if branch is omitted

  • repo_id (Optional[str]) -- optional ID of existing repository. Should be specified if repo_path is omitted

  • repo_path (Optional[str]) -- optional path of existing repository. Should be specified if repo_id is omitted

  • databricks_conn_id (str) -- Reference to the Databricks connection. By default and in the common case this will be databricks_default. To use token based authentication, provide the key token in the extra field for the connection and create the key host and leave the host field empty.

  • databricks_retry_limit (int) -- Amount of times retry if the Databricks backend is unreachable. Its value must be greater than or equal to 1.

  • databricks_retry_delay (int) -- Number of seconds to wait between retries (it might be a floating point number).

template_fields :Sequence[str] = ['repo_path', 'tag', 'branch'][source]
execute(self, context)[source]

This is the main method to derive when creating an operator. Context is the same dictionary used as when rendering jinja templates.

Refer to get_template_context for more context.

Was this entry helpful?