Remote Execution overview

This is feature is only available if you are on the Enterprise tier or above. See Astro Plans and Pricing.
Airflow 3
This feature is only available for Airflow 3.x Deployments.

Remote Execution is an execution mode on Astro that separates task execution from orchestration. With Remote Execution, you run Airflow tasks in your own Kubernetes infrastructure while Astro manages the orchestration components in the cloud. Use Remote Execution to keep your data and code within your own infrastructure for security or compliance requirements.

Astro Remote architecture overview

How Remote Execution works

Remote Execution uses a decoupled architecture with two planes:

Orchestration plane (Astro-managed)

The orchestration plane runs in Astro’s cloud infrastructure and includes:

  • Scheduler: Determines when dags and tasks should run
  • Web Server/API Server: Provides the Airflow UI and REST API
  • Metadata Database: Stores Dag and task metadata
  • Remote Execution API: Manages agent communication and task distribution

The orchestration plane assigns tasks to agents, monitors their health via heartbeats, and provides visibility through logging and observability features.

Execution plane (Customer-managed)

The execution plane runs in your Kubernetes cluster and includes:

  • Remote Execution Agents: Deployed via Helm charts, each agent includes:
    • Dag Processor: Parses and serializes Dag code
    • Triggerer: Manages deferrable tasks
    • Worker: Executes Airflow tasks
    • Sentinel: (Optional, recommended) Monitors agent health and performance

Agents pull tasks from the orchestration plane via secure HTTPS connections, execute them locally, and report status back. Agents maintain frequent heartbeats with the API Server. If an agent loses its heartbeat, the orchestration plane automatically reroutes tasks to healthy agents. See Remote Execution Agent failure scenarios for more information.

Key concepts

Remote Execution Agents

Agents are the core component of Remote Execution. Each agent is a collection of Airflow components (Worker, Dag Processor, Triggerer) deployed as a single unit in your Kubernetes cluster. You can deploy multiple agents with unique configurations across different clusters, regions, or node types to meet your workload requirements.

Agents communicate with Astro using:

  • Agent tokens: Authenticate agents to the orchestration plane
  • Outbound-only connections: Enable communication from your infrastructure to Astro without requiring inbound traffic
  • Heartbeat mechanism: Monitor agent health with regular status checks

Dag bundles

Dag bundles are collections of Dag files and supporting code. Remote Execution supports two types:

  • GitDagBundle: Dags stored in a Git repository (recommended for production)
  • LocalDagBundle: Dags stored in the container image or persistent volume

GitDagBundle provides automatic Dag versioning, allowing you to track changes and view different versions in the Airflow UI.

XCom backend

XCom (cross-communication) allows Airflow tasks to share data. With Remote Execution, you must configure an object storage backend (AWS S3, Azure Blob Storage, or GCP Cloud Storage) to pass XCom data between tasks running on different agents.

Secrets backend

Remote Execution Agents must be configured with a secrets backend to securely access Airflow connections and variables. This can be AWS Secrets Manager, Azure Key Vault, Google Cloud Secret Manager, or HashiCorp Vault.

What you need to configure

Configure the following required and optional components for Remote Execution.

Required components

Optional components

  • Sentinel (recommended): Monitor agent health and performance
  • Logging: Export task logs to external logging platforms or object storage
  • OpenLineage: Enable data lineage and observability features

Next steps