airflow.contrib.operators.oracle_to_azure_data_lake_transfer

Module Contents

class airflow.contrib.operators.oracle_to_azure_data_lake_transfer.OracleToAzureDataLakeTransfer(filename, azure_data_lake_conn_id, azure_data_lake_path, oracle_conn_id, sql, sql_params=None, delimiter=',', encoding='utf-8', quotechar='"', quoting=csv.QUOTE_MINIMAL, *args, **kwargs)[source]

Bases: airflow.models.BaseOperator

Moves data from Oracle to Azure Data Lake. The operator runs the query against Oracle and stores the file locally before loading it into Azure Data Lake.

Parameters
  • filename (str) – file name to be used by the csv file.

  • azure_data_lake_conn_id (str) – destination azure data lake connection.

  • azure_data_lake_path (str) – destination path in azure data lake to put the file.

  • oracle_conn_id (str) – source Oracle connection.

  • sql (str) – SQL query to execute against the Oracle database. (templated)

  • sql_params (str) – Parameters to use in sql query. (templated)

  • delimiter (str) – field delimiter in the file.

  • encoding (str) – encoding type for the file.

  • quotechar (str) – Character to use in quoting.

  • quoting (str) – Quoting strategy. See unicodecsv quoting for more information.

template_fields = ['filename', 'sql', 'sql_params'][source]
ui_color = #e08c8c[source]
_write_temp_file(self, cursor, path_to_save)[source]
execute(self, context)[source]

Was this entry helpful?