Content
Command Line Interface Reference¶
Airflow has a very rich command line interface that allows for many types of operation on a DAG, starting services, and supporting development and testing.
usage: airflow [-h] GROUP_OR_COMMAND ...
Positional Arguments¶
- GROUP_OR_COMMAND
- Possible choices: celery, config, connections, dags, db, kerberos, kubernetes, pools, rotate-fernet-key, sync-perm, tasks, users, variables, backfill, generate_pod_template, serve_logs, scheduler, webserver, version, info 
Sub-commands:¶
celery¶
Start celery components. Works only when using CeleryExecutor. For more information, see https://airflow.apache.org/docs/stable/executor/celery.html
airflow celery [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: flower, worker 
Sub-commands:¶
flower¶
Start a Celery Flower
airflow celery flower [-h] [-A BASIC_AUTH] [-a BROKER_API] [-D]
                      [-c FLOWER_CONF] [-H HOSTNAME] [-l LOG_FILE]
                      [--pid [PID]] [-p PORT] [--stderr STDERR]
                      [--stdout STDOUT] [-u URL_PREFIX]
Named Arguments¶
- -A, --basic-auth
- Securing Flower with Basic Authentication. Accepts user:password pairs separated by a comma. Example: flower_basic_auth = user1:password1,user2:password2 - Default: “” 
- -a, --broker-api
- Broker API 
- -D, --daemon
- Daemonize instead of running in the foreground - Default: False 
- -c, --flower-conf
- Configuration file for flower 
- -H, --hostname
- Set the hostname on which to run the server - Default: “0.0.0.0” 
- -l, --log-file
- Location of the log file 
- --pid
- PID file location 
- -p, --port
- The port on which to run the server - Default: 5555 
- --stderr
- Redirect stderr to this file 
- --stdout
- Redirect stdout to this file 
- -u, --url-prefix
- URL prefix for Flower - Default: “” 
worker¶
Start a Celery worker node
airflow celery worker [-h] [-a AUTOSCALE] [-H CELERY_HOSTNAME]
                      [-c CONCURRENCY] [-D] [-l LOG_FILE] [--pid [PID]]
                      [-q QUEUES] [-s] [--stderr STDERR] [--stdout STDOUT]
Named Arguments¶
- -a, --autoscale
- Minimum and Maximum number of worker to autoscale 
- -H, --celery-hostname
- Set the hostname of celery worker if you have multiple workers on a single machine 
- -c, --concurrency
- The number of worker processes - Default: 16 
- -D, --daemon
- Daemonize instead of running in the foreground - Default: False 
- -l, --log-file
- Location of the log file 
- --pid
- PID file location 
- -q, --queues
- Comma delimited list of queues to serve - Default: “default” 
- -s, --skip-serve-logs
- Don’t start the serve logs process along with the workers - Default: False 
- --stderr
- Redirect stderr to this file 
- --stdout
- Redirect stdout to this file 
config¶
View configuration
airflow config [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: list 
connections¶
Manage connections
airflow connections [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: add, delete, list 
Sub-commands:¶
add¶
Add a connection
airflow connections add [-h] [--conn-extra CONN_EXTRA] [--conn-host CONN_HOST]
                        [--conn-login CONN_LOGIN]
                        [--conn-password CONN_PASSWORD]
                        [--conn-port CONN_PORT] [--conn-schema CONN_SCHEMA]
                        [--conn-type CONN_TYPE] [--conn-uri CONN_URI]
                        conn_id
Positional Arguments¶
- conn_id
- Connection id, required to get/add/delete a connection 
Named Arguments¶
- --conn-extra
- Connection Extra field, optional when adding a connection 
- --conn-host
- Connection host, optional when adding a connection 
- --conn-login
- Connection login, optional when adding a connection 
- --conn-password
- Connection password, optional when adding a connection 
- --conn-port
- Connection port, optional when adding a connection 
- --conn-schema
- Connection schema, optional when adding a connection 
- --conn-type
- Connection type, required to add a connection without conn_uri 
- --conn-uri
- Connection URI, required to add a connection without conn_type 
dags¶
Manage DAGs
airflow dags [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: delete, list, list-runs, next-execution, pause, report, show, state, trigger, unpause 
Sub-commands:¶
delete¶
Delete all DB records related to the specified DAG
airflow dags delete [-h] [-y] dag_id
Positional Arguments¶
- dag_id
- The id of the dag 
Named Arguments¶
- -y, --yes
- Do not prompt to confirm reset. Use with care! - Default: False 
list¶
List all the DAGs
airflow dags list [-h] [-S SUBDIR]
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
list-runs¶
List DAG runs given a DAG id. If state option is given, it will only search for all the dagruns with the given state. If no_backfill option is given, it will filter out all backfill dagruns for given dag id. If start_date is given, it will filter out all the dagruns that were executed before this date. If end_date is given, it will filter out all the dagruns that were executed after this date.
airflow dags list-runs [-h] [-d DAG_ID] [--no-backfill] [--state STATE]
Named Arguments¶
- -d, --dag-id
- The id of the dag 
- --no-backfill
- filter all the backfill dagruns given the dag id - Default: False 
- --state
- Only list the dag runs corresponding to the state 
next-execution¶
Get the next execution datetimes of a DAG. It returns one execution unless the num-executions option is given
airflow dags next-execution [-h] [-S SUBDIR] dag_id
Positional Arguments¶
- dag_id
- The id of the dag 
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
pause¶
Pause a DAG
airflow dags pause [-h] [-S SUBDIR] dag_id
Positional Arguments¶
- dag_id
- The id of the dag 
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
report¶
Show DagBag loading report
airflow dags report [-h] [-S SUBDIR]
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
show¶
The –imgcat option only works in iTerm.
For more information, see: https://www.iterm2.com/documentation-images.html
The –save option saves the result to the indicated file.
The file format is determined by the file extension. For more information about supported format, see: https://www.graphviz.org/doc/info/output.html
If you want to create a PNG file then you should execute the following command: airflow dags show <DAG_ID> –save output.png
If you want to create a DOT file then you should execute the following command: airflow dags show <DAG_ID> –save output.dot
airflow dags show [-h] [--imgcat] [-s SAVE] [-S SUBDIR] dag_id
Positional Arguments¶
- dag_id
- The id of the dag 
Named Arguments¶
- --imgcat
- Displays graph using the imgcat tool. - Default: False 
- -s, --save
- Saves the result to the indicated file. 
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
state¶
Get the status of a dag run
airflow dags state [-h] [-S SUBDIR] dag_id execution_date
Positional Arguments¶
- dag_id
- The id of the dag 
- execution_date
- The execution date of the DAG 
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
trigger¶
Trigger a DAG run
airflow dags trigger [-h] [-c CONF] [-e EXEC_DATE] [-r RUN_ID] [-S SUBDIR]
                     dag_id
Positional Arguments¶
- dag_id
- The id of the dag 
Named Arguments¶
- -c, --conf
- JSON string that gets pickled into the DagRun’s conf attribute 
- -e, --exec-date
- The execution date of the DAG 
- -r, --run-id
- Helps to identify this run 
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
unpause¶
Resume a paused DAG
airflow dags unpause [-h] [-S SUBDIR] dag_id
Positional Arguments¶
- dag_id
- The id of the dag 
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
db¶
Database operations
airflow db [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: check, init, reset, shell, upgrade 
kerberos¶
Start a kerberos ticket renewer
airflow kerberos [-h] [-D] [-k [KEYTAB]] [-l LOG_FILE] [--pid [PID]]
                 [--stderr STDERR] [--stdout STDOUT]
                 [principal]
Positional Arguments¶
- principal
- kerberos principal 
Named Arguments¶
- -D, --daemon
- Daemonize instead of running in the foreground - Default: False 
- -k, --keytab
- keytab - Default: “airflow.keytab” 
- -l, --log-file
- Location of the log file 
- --pid
- PID file location 
- --stderr
- Redirect stderr to this file 
- --stdout
- Redirect stdout to this file 
kubernetes¶
Tools to help run the KubernetesExecutor
airflow kubernetes [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: cleanup-pods, generate-dag-yaml 
Sub-commands:¶
cleanup-pods¶
Clean up Kubernetes pods in evicted/failed/succeeded states
airflow kubernetes cleanup-pods [-h] [--namespace NAMESPACE]
Named Arguments¶
- --namespace
- Kubernetes Namespace - Default: “default” 
generate-dag-yaml¶
Generate YAML files for all tasks in DAG. Useful for debugging tasks without launching into a cluster
airflow kubernetes generate-dag-yaml [-h] [-o OUTPUT_PATH] [-S SUBDIR]
                                     dag_id execution_date
Positional Arguments¶
- dag_id
- The id of the dag 
- execution_date
- The execution date of the DAG 
Named Arguments¶
- -o, --output-path
- The output for generated yaml files - Default: “[CWD]” 
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
pools¶
Manage pools
airflow pools [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: delete, export, get, import, list, set 
rotate-fernet-key¶
Rotate all encrypted connection credentials and variables; see https://airflow.apache.org/docs/stable/howto/secure-connections.html#rotating-encryption-keys
airflow rotate-fernet-key [-h]
tasks¶
Manage tasks
airflow tasks [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: clear, failed-deps, list, render, run, state, test 
Sub-commands:¶
clear¶
Clear a set of task instance, as if they never ran
airflow tasks clear [-h] [-R] [-d] [-e END_DATE] [-X] [-x] [-f] [-r]
                    [-s START_DATE] [-S SUBDIR] [-t TASK_REGEX] [-u] [-y]
                    dag_id
Positional Arguments¶
- dag_id
- The id of the dag 
Named Arguments¶
- -R, --dag-regex
- Search dag_id as regex instead of exact string - Default: False 
- -d, --downstream
- Include downstream tasks - Default: False 
- -e, --end-date
- Override end_date YYYY-MM-DD 
- -X, --exclude-parentdag
- Exclude ParentDAGS if the task cleared is a part of a SubDAG - Default: False 
- -x, --exclude-subdags
- Exclude subdags - Default: False 
- -f, --only-failed
- Only failed jobs - Default: False 
- -r, --only-running
- Only running jobs - Default: False 
- -s, --start-date
- Override start_date YYYY-MM-DD 
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
- -t, --task-regex
- The regex to filter specific task_ids to backfill (optional) 
- -u, --upstream
- Include upstream tasks - Default: False 
- -y, --yes
- Do not prompt to confirm reset. Use with care! - Default: False 
failed-deps¶
Returns the unmet dependencies for a task instance from the perspective of the scheduler. In other words, why a task instance doesn’t get scheduled and then queued by the scheduler, and then run by an executor.
airflow tasks failed-deps [-h] [-S SUBDIR] dag_id task_id execution_date
Positional Arguments¶
- dag_id
- The id of the dag 
- task_id
- The id of the task 
- execution_date
- The execution date of the DAG 
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
list¶
List the tasks within a DAG
airflow tasks list [-h] [-S SUBDIR] [-t] dag_id
Positional Arguments¶
- dag_id
- The id of the dag 
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
- -t, --tree
- Tree view - Default: False 
render¶
Render a task instance’s template(s)
airflow tasks render [-h] [-S SUBDIR] dag_id task_id execution_date
Positional Arguments¶
- dag_id
- The id of the dag 
- task_id
- The id of the task 
- execution_date
- The execution date of the DAG 
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
run¶
Run a single task instance
airflow tasks run [-h] [--cfg-path CFG_PATH] [-f] [-A] [-i] [-I] [-N] [-l]
                  [-m] [-p PICKLE] [--pool POOL] [--ship-dag] [-S SUBDIR]
                  dag_id task_id execution_date
Positional Arguments¶
- dag_id
- The id of the dag 
- task_id
- The id of the task 
- execution_date
- The execution date of the DAG 
Named Arguments¶
- --cfg-path
- Path to config file to use instead of airflow.cfg 
- -f, --force
- Ignore previous task instance state, rerun regardless if task already succeeded/failed - Default: False 
- -A, --ignore-all-dependencies
- Ignores all non-critical dependencies, including ignore_ti_state and ignore_task_deps - Default: False 
- -i, --ignore-dependencies
- Ignore task-specific dependencies, e.g. upstream, depends_on_past, and retry delay dependencies - Default: False 
- -I, --ignore-depends-on-past
- Ignore depends_on_past dependencies (but respect upstream dependencies) - Default: False 
- -N, --interactive
- Do not capture standard output and error streams (useful for interactive debugging) - Default: False 
- -l, --local
- Run the task using the LocalExecutor - Default: False 
- -m, --mark-success
- Mark jobs as succeeded without running them - Default: False 
- -p, --pickle
- Serialized pickle object of the entire dag (used internally) 
- --pool
- Resource pool to use 
- --ship-dag
- Pickles (serializes) the DAG and ships it to the worker - Default: False 
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
state¶
Get the status of a task instance
airflow tasks state [-h] [-S SUBDIR] dag_id task_id execution_date
Positional Arguments¶
- dag_id
- The id of the dag 
- task_id
- The id of the task 
- execution_date
- The execution date of the DAG 
Named Arguments¶
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
test¶
Test a task instance. This will run a task without checking for dependencies or recording its state in the database
airflow tasks test [-h] [-n] [-m] [-S SUBDIR] [-t TASK_PARAMS]
                   dag_id task_id execution_date
Positional Arguments¶
- dag_id
- The id of the dag 
- task_id
- The id of the task 
- execution_date
- The execution date of the DAG 
Named Arguments¶
- -n, --dry-run
- Perform a dry run for each task. Only renders Template Fields for each task, nothing else - Default: False 
- -m, --post-mortem
- Open debugger on uncaught exception - Default: False 
- -S, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
- -t, --task-params
- Sends a JSON params dict to the task 
users¶
Manage users
airflow users [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: create, delete, list 
Sub-commands:¶
create¶
Create a user
airflow users create [-h] -e EMAIL -f FIRSTNAME -l LASTNAME [-p PASSWORD] -r
                     ROLE [--use-random-password] -u USERNAME
Named Arguments¶
- -e, --email
- Email of the user 
- -f, --firstname
- First name of the user 
- -l, --lastname
- Last name of the user 
- -p, --password
- Password of the user, required to create a user without –use-random-password 
- -r, --role
- Role of the user. Existing roles include Admin, User, Op, Viewer, and Public 
- --use-random-password
- Do not prompt for password. Use random string instead. Required to create a user without –password - Default: False 
- -u, --username
- Username of the user 
examples: To create an user with “Admin” role and username equals to “admin”, run:
- $ airflow users create
–username admin –firstname FIRST_NAME –lastname LAST_NAME –role Admin –email admin@example.org
variables¶
Manage variables
airflow variables [-h] COMMAND ...
Positional Arguments¶
- COMMAND
- Possible choices: delete, export, get, import, set 
Sub-commands:¶
export¶
Export all variables
airflow variables export [-h] file
Positional Arguments¶
- file
- Export all variables to JSON file 
get¶
Get variable
airflow variables get [-h] [-d VAL] [-j] key
Positional Arguments¶
- key
- Variable key 
Named Arguments¶
- -d, --default
- Default value returned if variable does not exist 
- -j, --json
- Deserialize JSON variable - Default: False 
backfill¶
Run subsections of a DAG for a specified date range. If reset_dag_run option is used, backfill will first prompt users whether airflow should clear all the previous dag_run and task_instances within the backfill date range. If rerun_failed_tasks is used, backfill will auto re-run the previous failed task instances within the backfill date range.
airflow backfill [-h] [-t TASK_REGEX] [-s START_DATE] [-e END_DATE] [-m] [-l]
                 [-x] [-y] [-i] [-I] [-sd SUBDIR] [--pool POOL]
                 [--delay_on_limit DELAY_ON_LIMIT] [-dr] [-v] [-c CONF]
                 [--reset_dagruns] [--rerun_failed_tasks] [-B]
                 dag_id
Positional Arguments¶
- dag_id
- The id of the dag 
Named Arguments¶
- -t, --task_regex
- The regex to filter specific task_ids to backfill (optional) 
- -s, --start_date
- Override start_date YYYY-MM-DD 
- -e, --end_date
- Override end_date YYYY-MM-DD 
- -m, --mark_success
- Mark jobs as succeeded without running them - Default: False 
- -l, --local
- Run the task using the LocalExecutor - Default: False 
- -x, --donot_pickle
- Do not attempt to pickle the DAG object to send over to the workers, just tell the workers to run their version of the code. - Default: False 
- -y, --yes
- Do not prompt to confirm reset. Use with care! - Default: False 
- -i, --ignore_dependencies
- Skip upstream tasks, run only the tasks matching the regexp. Only works in conjunction with task_regex - Default: False 
- -I, --ignore_first_depends_on_past
- Ignores depends_on_past dependencies for the first set of tasks only (subsequent executions in the backfill DO respect depends_on_past). - Default: False 
- -sd, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
- --pool
- Resource pool to use 
- --delay_on_limit
- Amount of time in seconds to wait when the limit on maximum active dag runs (max_active_runs) has been reached before trying to execute a dag run again. - Default: 1.0 
- -dr, --dry_run
- Perform a dry run for each task. Only renders Template Fields for each task, nothing else - Default: False 
- -v, --verbose
- Make logging output more verbose - Default: False 
- -c, --conf
- JSON string that gets pickled into the DagRun’s conf attribute 
- --reset_dagruns
- if set, the backfill will delete existing backfill-related DAG runs and start anew with fresh, running DAG runs - Default: False 
- --rerun_failed_tasks
- if set, the backfill will auto-rerun all the failed tasks for the backfill date range instead of throwing exceptions - Default: False 
- -B, --run_backwards
- if set, the backfill will run tasks from the most recent day first. if there are tasks that depend_on_past this option will throw an exception - Default: False 
generate_pod_template¶
Reads your airflow.cfg and migrates your configurations into a airflow_template.yaml file. From this point a user can linkthis file to airflow using the pod_template_file argumentand modify using the Kubernetes API
airflow generate_pod_template [-h] [-o OUTPUT_PATH]
Named Arguments¶
- -o, --output-path
- output path for yaml file - Default: “/Users/kamilbregula/devel/google-airflow/airflow/docs” 
scheduler¶
Start a scheduler instance
airflow scheduler [-h] [-d DAG_ID] [-sd SUBDIR] [-r RUN_DURATION]
                  [-n NUM_RUNS] [-p] [--pid [PID]] [-D] [--stdout STDOUT]
                  [--stderr STDERR] [-l LOG_FILE]
Named Arguments¶
- -d, --dag_id
- The id of the dag to run 
- -sd, --subdir
- File location or directory from which to look for the dag. Defaults to ‘[AIRFLOW_HOME]/dags’ where [AIRFLOW_HOME] is the value you set for ‘AIRFLOW_HOME’ config you set in ‘airflow.cfg’ - Default: “[AIRFLOW_HOME]/dags” 
- -r, --run-duration
- Set number of seconds to execute before exiting 
- -n, --num_runs
- Set the number of runs to execute before exiting - Default: -1 
- -p, --do_pickle
- Attempt to pickle the DAG object to send over to the workers, instead of letting workers run their version of the code. - Default: False 
- --pid
- PID file location 
- -D, --daemon
- Daemonize instead of running in the foreground - Default: False 
- --stdout
- Redirect stdout to this file 
- --stderr
- Redirect stderr to this file 
- -l, --log-file
- Location of the log file 
webserver¶
Start a Airflow webserver instance
airflow webserver [-h] [-p PORT] [-w WORKERS]
                  [-k {sync,eventlet,gevent,tornado}] [-t WORKER_TIMEOUT]
                  [-hn HOSTNAME] [--pid [PID]] [-D] [--stdout STDOUT]
                  [--stderr STDERR] [-A ACCESS_LOGFILE] [-E ERROR_LOGFILE]
                  [-l LOG_FILE] [--ssl_cert SSL_CERT] [--ssl_key SSL_KEY] [-d]
Named Arguments¶
- -p, --port
- The port on which to run the server - Default: 8080 
- -w, --workers
- Number of workers to run the webserver on - Default: 4 
- -k, --workerclass
- Possible choices: sync, eventlet, gevent, tornado - The worker class to use for Gunicorn - Default: “sync” 
- -t, --worker_timeout
- The timeout for waiting on webserver workers - Default: 120 
- -hn, --hostname
- Set the hostname on which to run the web server - Default: “0.0.0.0” 
- --pid
- PID file location 
- -D, --daemon
- Daemonize instead of running in the foreground - Default: False 
- --stdout
- Redirect stdout to this file 
- --stderr
- Redirect stderr to this file 
- -A, --access_logfile
- The logfile to store the webserver access log. Use ‘-‘ to print to stderr. - Default: “-“ 
- -E, --error_logfile
- The logfile to store the webserver error log. Use ‘-‘ to print to stderr. - Default: “-“ 
- -l, --log-file
- Location of the log file 
- --ssl_cert
- Path to the SSL certificate for the webserver - Default: “” 
- --ssl_key
- Path to the key to use with the SSL certificate - Default: “” 
- -d, --debug
- Use the server that ships with Flask in debug mode - Default: False 
info¶
Show information about current Airflow and environment
airflow info [-h] [--anonymize] [--file-io]
Named Arguments¶
- --anonymize
- Minimize any personal identifiable information. Use it when sharing output with others. - Default: False 
- --file-io
- Send output to file.io service and returns link. - Default: False