Skip to content

Using Apache Airflow to Create Dynamic, Extensible Data Workflows on GCP

Photo of Remy Rosenbaum
Hosted By
Remy R. and Joe C.
Using Apache Airflow to Create Dynamic, Extensible Data Workflows on GCP

Details

>> Can't make it? Register for our live webinar version on May 10th at 2pm ET: https://goo.gl/hC46sw
_____________________________________________________________________

Join SoulCycle, Caserta and more than 150 fellow data nerds for pizza, drinks, mingling and presentations. Data architects, data engineers and data scientists alike will enjoy the contents of the talks.

TOPIC:
Using Apache Airflow to create dynamic, extensible, elegant, scalable data workflows on Google Cloud at SoulCycle.

Caserta consultants will share, in technical detail, how to define and configure Airflow pipelines to create workflows that are maintainable, versionable, testable, and collaborative.

In this meetup you will learn best practices to create Airflow directed acyclic graphs (DAGs) of tasks, enforcing relationships and dependencies. The talk covers constructing Airflow pipelines with Spark based ETL, a Google Storage Data Lake, and a Big Query Data Warehouse.

The stage will be shared with Dallas S. Simpson, Director of Data Engineering at SoulCycle Inc. Dallas will co-present with Caserta to share his experience and lessons learned using Airflow on their internal GCP Data Analytics Platform project.

Photo of Big Data Warehousing group
Big Data Warehousing
See more events
SoulCycle HQ
609 Greenwich Street · New York, ny