CLI Quickstart

Astronomer's open-source CLI is the easiest way to run Apache Airflow on your machine.

From the CLI, you can establish a local testing environment and deploy to Astronomer, whether Cloud or Enterprise, whenever you're ready.


To start using the CLI, make sure you've already installed:

  • Docker, version 18.09 or higher


If you're a Cloud customer, run:

$ curl -sSL | sudo bash -s -- v0.7.5-2

If you are an Enterprise customer, run:

$ curl -sSL | sudo bash

Note: The curl command will work for Unix (Linux+Mac) based systems. If you want to run on Windows 10, you'll need to run through this guide on getting Docker for WSL working.

Confirm the Install

Let's make sure you have Astro CLI installed on your machine, and that you have a project to work from.

$ astro

If you're set up properly, you should see the following:

astro is a command line interface for working with the Astronomer Platform.

  astro [command]

Available Commands:
  airflow     Manage airflow projects and deployments
  auth        Mangage astronomer identity
  cluster     Manage Astronomer EE clusters
  config      Manage astro project configurations
  deployment  Manage airflow deployments
  help        Help about any command
  upgrade     Check for newer version of Astronomer CLI
  user        Manage astronomer user
  version     Astronomer CLI version
  workspace   Manage Astronomer workspaces

  -h, --help   help for astro

For a breakdown of subcommands and corresponding descriptions, you can run: astro help

Create a Project

Your first step is to create a project to work from that lives in a folder on your local machine. The command you'll need is listed below, with an example hello-astro project.

mkdir hello-astro && cd hello-astro
astro airflow init

astro airflow init will build a base image from Astronomer's fork of Apache-Airflow using Alpine Linux. The build process will include everything in your project directory, which makes it easy to include any shell scripts, static files, or anything else you want to include in your code.

Once that command is run, you'll see the following skeleton project generated:

├── dags # Where your DAGs go
│   ├──
├── Dockerfile # For runtime overrides
├── include # For any other files you'd like to include
├── packages.txt # For OS-level packages
├── plugins # For any custom or community Airflow plugins
└── requirements.txt # For any python packages

The sample dag that gets generated has a few tasks that run bash commands

Note: The image will take some time to build the first time, after that it will build from cached layers.

Now you can run astro airflow start and see Airflow running on localhost:8080/admin

All changes made to the dags and plugins directory will be picked up automatically - any changes made to any of the other files will need the image to be rebuilt (astro airflow stop and astro airflow start).

For more information on using the CLI, see the Developing Locally with the CLI section.