Astronomer's the Dataflow Cast

Why Airflow Became the Scheduling Backbone at Condé Nast Technology Lab with Arun Karthik

Data platforms are moving from batch-first pipelines to near real-time systems where orchestration, observability, scalability and governance all have to work together.

In this episode, Arun Karthik, Director, Data Solutions Engineering at Condé Nast Technology Lab, joins us to share how data engineering evolves from relational databases and ETL into distributed processing, modern orchestration with Apache Airflow and managed Airflow with Astronomer.

Key Takeaways:

  • 00:00 Introduction.
  • 02:13 Early data systems rely heavily on relational databases and batch-oriented processing models.
  • 07:01 Scheduling requirements evolve beyond fixed time windows as dependencies increase.
  • 10:14 Ease of use and developer experience influence adoption of orchestration frameworks.
  • 13:22 Operating open source orchestration tools requires ongoing engineering effort.
  • 14:45 Managed services help teams reduce infrastructure and maintenance responsibilities.
  • 17:27 Observability improves confidence in pipeline execution and system health.
  • 19:12 Governance considerations grow in importance as data platforms mature.
  • 20:46 Building data systems requires balancing speed, reliability and long-term sustainability.

Resources Mentioned:

Thanks for listening to “The Data Flowcast: Mastering Apache Airflow® for Data Engineering and AI.” If you enjoyed this episode, please leave a 5-star review to help get the word out about the show. And be sure to subscribe so you never miss any of the insightful conversations.

Be Our Guest

Interested in being a guest on The Data Flowcast? Fill out the form and we will be in touch.

Build, run, & observe your data workflows.
All in one place.

Build, run, & observe
your data workflows.
All in one place.

Try Astro today and get up to $20 in free credits during your 14-day trial.