SQLExecuteQueryOperator to connect to Snowflake

Use the SQLExecuteQueryOperator to execute SQL commands in a Snowflake database.

Warning

Previously, SnowflakeOperator was used to perform this kind of operation. But at the moment SnowflakeOperator is deprecated and will be removed in future versions of the provider. Please consider to switch to SQLExecuteQueryOperator as soon as possible.

Using the Operator

Use the conn_id argument to connect to your Snowflake instance where the connection metadata is structured as follows:

Snowflake Airflow Connection Metadata

Parameter

Input

Login: string

Snowflake user name

Password: string

Password for Snowflake user

Schema: string

Set schema to execute SQL operations on by default

Extra: dictionary

warehouse, account, database, region, role, authenticator

An example usage of the SQLExecuteQueryOperator to connect to Snowflake is as follows:

tests/system/providers/snowflake/example_snowflake.py[source]

snowflake_op_sql_str = SQLExecuteQueryOperator(
    task_id="snowflake_op_sql_str", sql=CREATE_TABLE_SQL_STRING
)

snowflake_op_with_params = SQLExecuteQueryOperator(
    task_id="snowflake_op_with_params",
    sql=SQL_INSERT_STATEMENT,
    parameters={"id": 56},
)

snowflake_op_sql_list = SQLExecuteQueryOperator(task_id="snowflake_op_sql_list", sql=SQL_LIST)

snowflake_op_sql_multiple_stmts = SQLExecuteQueryOperator(
    task_id="snowflake_op_sql_multiple_stmts",
    sql=SQL_MULTIPLE_STMTS,
    split_statements=True,
)

snowflake_op_template_file = SQLExecuteQueryOperator(
    task_id="snowflake_op_template_file",
    sql="example_snowflake_snowflake_op_template_file.sql",
)

Note

Parameters that can be passed onto the operator will be given priority over the parameters already given in the Airflow connection metadata (such as schema, role, database and so forth).

SnowflakeSqlApiOperator

Use the SnowflakeSqlApiHook to execute SQL commands in a Snowflake database.

You can also run this operator in deferrable mode by setting deferrable param to True. This will ensure that the task is deferred from the Airflow worker slot and polling for the task status happens on the trigger.

Using the Operator

Use the snowflake_conn_id argument to connect to your Snowflake instance where the connection metadata is structured as follows:

Snowflake Airflow Connection Metadata

Parameter

Input

Login: string

Snowflake user name. If using OAuth connection this is the client_id

Password: string

Password for Snowflake user. If using OAuth this is the client_secret

Schema: string

Set schema to execute SQL operations on by default

Extra: dictionary

warehouse, account, database, region, role, authenticator, refresh_token. If using OAuth must specify refresh_token (obtained here)

An example usage of the SnowflakeSqlApiHook is as follows:

tests/system/providers/snowflake/example_snowflake.py[source]

snowflake_sql_api_op_sql_multiple_stmt = SnowflakeSqlApiOperator(
    task_id="snowflake_op_sql_multiple_stmt",
    sql=SQL_MULTIPLE_STMTS,
    statement_count=len(SQL_LIST),
)

Note

Parameters that can be passed onto the operator will be given priority over the parameters already given in the Airflow connection metadata (such as schema, role, database and so forth).

Was this entry helpful?