DAG Serialization

In order to make Airflow Webserver stateless, Airflow >=1.10.7 supports DAG Serialization and DB Persistence. From Airflow 2.0.0, the Scheduler also uses Serialized DAGs for consistency and makes scheduling decisions.

_images/dag_serialization.png

Without DAG Serialization & persistence in DB, the Webserver and the Scheduler both need access to the DAG files. Both the Scheduler and Webserver parse the DAG files.

With DAG Serialization we aim to decouple the Webserver from DAG parsing which would make the Webserver very light-weight.

As shown in the image above, when using this feature, the DagFileProcessorProcess in the Scheduler parses the DAG files, serializes them in JSON format and saves them in the Metadata DB as SerializedDagModel model.

The Webserver now instead of having to parse the DAG files again, reads the serialized DAGs in JSON, de-serializes them and creates the DagBag and uses it to show in the UI. And the Scheduler does not need the actual DAGs for making scheduling decisions, instead of using the DAG files, we use the serialized DAGs that contain all the information needed to schedule the DAGs from Airflow 2.0.0 (this was done as part of Scheduler HA).

One of the key features that is implemented as a part of DAG Serialization is that instead of loading an entire DagBag when the WebServer starts we only load each DAG on demand from the Serialized Dag table. It helps to reduce the Webserver startup time and memory. This reduction is notable when you have a large number of DAGs.

You can enable the source code to be stored in the database to make the Webserver completely independent of the DAG files. This is not necessary if your files are embedded in the Docker image or you can otherwise provide them to the Webserver. The data is stored in the DagCode model.

The last element is rendering template fields. When serialization is enabled, templates are not rendered to requests, but a copy of the field contents is saved before the task is executed on worker. The data is stored in the RenderedTaskInstanceFields model. To limit the excessive growth of the database, only the most recent entries are kept and older entries are purged.

Note

DAG Serialization is strictly required and can not be turned off from Airflow 2.0+.

Dag Serialization Settings

Add the following settings in airflow.cfg:

[core]

# You can also update the following default configurations based on your needs
min_serialized_dag_update_interval = 30
min_serialized_dag_fetch_interval = 10
max_num_rendered_ti_fields_per_task = 30
compress_serialized_dags = False
  • min_serialized_dag_update_interval: This flag sets the minimum interval (in seconds) after which the serialized DAGs in the DB should be updated. This helps in reducing database write rate.

  • min_serialized_dag_fetch_interval: This option controls how often the Serialized DAG will be re-fetched from the DB when it is already loaded in the DagBag in the Webserver. Setting this higher will reduce load on the DB, but at the expense of displaying a possibly stale cached version of the DAG.

  • max_num_rendered_ti_fields_per_task: This option controls the maximum number of Rendered Task Instance Fields (Template Fields) per task to store in the Database.

  • compress_serialized_dags: This option controls whether to compress the Serialized DAG to the Database. It is useful when there are very large DAGs in your cluster. When True, this will disable the DAG dependencies view.

If you are updating Airflow from <1.10.7, please do not forget to run airflow db upgrade.

Limitations

  • When using user-defined filters and macros, the Rendered View in the Webserver might show incorrect results for TIs that have not yet executed as it might be using external modules that the Webserver won’t have access to. Use airflow tasks render CLI command in such situation to debug or test rendering of your template_fields. Once the tasks execution starts the Rendered Template Fields will be stored in the DB in a separate table and after which the correct values would be showed in the Webserver (Rendered View tab).

Note

You need Airflow >= 1.10.10 for completely stateless Webserver. Airflow 1.10.7 to 1.10.9 needed access to Dag files in some cases. More Information: https://airflow.apache.org/docs/1.10.9/dag-serialization.html#limitations

Using a different JSON Library

To use a different JSON library instead of the standard json library like ujson, you need to define a json variable in local Airflow settings (airflow_local_settings.py) file as follows:

import ujson

json = ujson

Was this entry helpful?