To perform some upfront analysis. Refer to the next section for more detail.
To know the following about your Airflow setup:
What operator types are used in the DAGs you're migrating
What Airflow connections your DAGs depend on
What Airflow variables you've set
What Airflow secrets backend you use
Where the permissions that your DAGs depend on are defined
If using Dagster Cloud, an existing Dagster Cloud account. While your migrated Airflow DAGs will work with Dagster Open Source, this guide includes setup specific to Dagster Cloud.
You may be coming to this document/library in a skeptical frame of mind, having previously been burned by projects that claim to have 100% foolproof, automated migration tools. We want to assure you that we are not making that claim.
The dagster-airflow migration library should be viewed as a powerful accelerant of migration, rather than guaranteeing completely automated migration. The amount of work required is proportional to the complexity of your installation. Smaller implementations of less complexity can be trivial to migrate, making it appear virtually automatic; larger, more complicated, or customized implementations require more investment.
For larger installations, teams that already adopt devops practices and/or have standardized usage of Airflow will have a smoother transition.
Some concrete examples:
If you rely on the usage of the Airflow UI to set production connections or variables, this will require a change of workflow as you will no longer have the Airflow UI at the end of this migration. If instead you rely on code-as-infrastructure patterns to set connections or variables, migration is more straightforward.
If you have standardized on a small set or even a single operator type (e.g. the K8sPodOperator), migration will be easier. If you use a large number of operator types with a wide range of infrastructure requirements, migration will be more work.
If you dynamically generate your Airflow DAGs from a higher-level API or DSL (e.g. yaml), the migration will be more straightforward than all your stakeholders directly creating Airflow DAGs.
Even in the case that requires some infrastructure investment, dagster-airflow dramatically eases migration, typically by orders of magnitude. The cost can be borne by a single or small set of infrastructure-oriented engineers, which dramatically reduces coordination costs. You do not have to move all of your stakeholders over to new APIs simultaneously. In our experience, practitioners welcome the change, because of the immediate improvement in tooling, stability, and development speed.
Step 1: Prepare your project for a new Dagster Python module#
While there are many ways to structure an Airflow git repository, this guide assumes you're using a repository structure that contains a single ./dags DagBag directory that contains all your DAGs.
In the root of your repository, create a dagster_migration.py file.
This step may require working through a number of version pins. Specifically, installing Airflow 1.x.x versions may be challenging due to (usually) outdated constraint files.
Don't get discouraged if you run into problems! Reach out to the Dagster Slack for help.
In this step, you'll install the dagster, dagster-airflow, and dagster-webserver Python packages alongside Airflow. We strongly recommend using a virtualenv.
We also suggest verifying that you're installing the correct versions of your Airflow dependencies. Verifying the dependency versions will likely save you from debugging tricky errors later.
To check dependency versions, open your Airflow provider's UI and locate the version numbers. When finished, continue to the next step.
In the dagster_migration.py file you created in Step 1, use make_dagster_definitions_from_airflow_dags_path and pass in the file path of your Airflow DagBag. Dagster will load the DagBag and convert all DAGs into Dagster jobs and schedules.
import os
from dagster_airflow import(
make_dagster_definitions_from_airflow_dags_path,)
migrated_airflow_definitions = make_dagster_definitions_from_airflow_dags_path(
os.path.abspath("./dags/"),)
In this step, you'll spin up Dagster's web-based UI, and verify that your migrated DAGs are loading. Note: Unless the migrated DAGs depend on no Airflow configuration state or permissions, it's unlikely they'll execute correctly at this point. That's okay - we'll fix it in a bit. Starting the Dagster UI is the first step in our development loop, allowing you to make a local change, view it in the UI, and debug any errors.
Run the following to start the UI:
dagster dev -f ./migrate_repo.py
In your browser, navigate to http://localhost:3001. You should see a list of Dagster jobs that correspond to the DAGs in your Airflow DagBag.
Run one of the simpler jobs, ideally one where you're familiar with the business logic. Note that it's likely to fail due to a configuration or permissions issue.
Using logs to identify and making configuration changes to fix the cause of the failure.
Repeat these steps as needed until the jobs run successfully.
There are a variety of Airflow Operator types that are used to launch compute in various external execution environments, for example Kubernetes or Amazon ECS. When getting things working locally we'd recommend trying to execute those containers locally unless it's either unrealistic or impossible to emulate the cloud environment. For example if you use the K8sPodOperator, it likely means that you will need to have a local Kubernetes cluster running, and in that case we recommend docker's built-in Kubernetes environment. You also need to be able to pull down the container images that will be needed for execution to your local machine.
If local execution is impossible, we recommend using Branch Deployments in Dagster Cloud, which is a well-supported workflow for cloud-native development.
To port your Airflow configuration, we recommend using environment variables as much as possible. Specifically, we recommend using a .env file containing Airflow variables and/or a secrets backend configuration in the root of your project.
You'll also need to configure the Airflow connections that your DAGs depend on. To accomplish this, use the connections parameter instead of URI-encoded environment variables.
import os
from airflow.models import Connection
from dagster_airflow import make_dagster_definitions_from_airflow_dags_path
migrated_airflow_definitions = make_dagster_definitions_from_airflow_dags_path(
os.path.abspath("./dags/"),
connections=[
Connection(conn_id="http_default", conn_type="uri", host="https://google.com")],)
Iterate as needed until all configuration is correctly ported to your local environment.
Step 6: Deciding on persistent vs ephemeral Airflow database#
The Dagster Airflow migration tooling supports two methods for persisting Airflow metadatabase state. By default, it will use an ephemeral database that is scoped to every job run and thrown away as soon as a job run terminates. You can also configure a persistent database that will be shared by all job runs. This tutorial uses the ephemeral database, but the persistent database option is recommended if you need the following features:
retry-from-failure support in Dagster
Passing Airflow state between DAG runs (i.e., xcoms)
SubDAGOperators
Airflow Pools
If you have complex Airflow DAGs (cross-DAG state sharing, pools) or the ability to retry-from-failure you will need to have a persistent database when making your migration.
You can configure your persistent Airflow database by providing an airflow_db to the resource_defs parameter of the dagster-airflow APIs:
This step is applicable to Dagster Cloud. If deploying to your infrastructure, refer to the Deployment guides for more info.
Additionally, until your Airflow DAGs execute successfully in your local environment, we recommend waiting to move to production.
In this step, you'll set up your project for use with Dagster Cloud.
Complete the steps in the Dagster Cloud Getting Started guide, if you haven't already. Proceed to the next step when your account is set up and you have the dagster-cloud CLI installed.
In the root of your project, create or modify the dagster_cloud.yaml file with the following code:
Your Airflow instance likely had specific IAM or Kubernetes permissions that allowed it to successfully run your Airflow DAGs. To run the migrated Dagster jobs, you'll need to duplicate these permissions for Dagster.
If you're unable to use Airflow connections or environment variables, you can attach permissions directly to the infrastructure where you're deploying Dagster.