search iconsearch icon
Type something to search...

Prefect Essentials: Basics, Setup and Migration

Prefect Essentials: Basics, Setup and Migration

0. Why Prefect?

Five years ago (in 2019), I created a pipeline to automate tasks (more info in Domain LogoVtasks). Initially, I used Domain LogoAirflow (see how in Domain LogoSetting up Airflow) since it was the industry standard at the time. However, I found Airflow to be overly complex for my needs. This led me to switch to Domain LogoLuigi (see how in Domain LogoLuigi orchestrator).

While Luigi worked better for me, I still lacked features such as dynamic DAG creation based on input parameters. Fortunately, I discovered Domain LogoPrefect, which offered the simplicity and flexibility I desired:

  • Super simple to start with
  • Very flexible

I quickly adopted Prefect for my pipeline and found it to be highly effective.

You might be hesitant to use a new tool without substantial support. However, Prefect is gaining popularity and is positioned as the top Airflow alternative, as evident from the GitHub stars of various orchestrators:

GitHub Stars extracted using Domain LogoDaily Stars explorer. To quickly see the star history of a repo, you can use Domain LogoStars History.

1. Prefect Basics

If you want a more in-depth introduction to Prefect basics, you can read Domain LogoPrefect Quickstart instead.

The first step is to install prefect with:

pip install prefect

1.1. First flow (aka DAG)

Now, you’re ready to create your first flow (this is how Prefect refers to a DAG) using a simple decorator:


from prefect import flow

def greet(name):
    print(f"Hello {name}")

if __name__ == '__main__':

To run the flow, execute the following command:


1.1. Adding tasks and subflows

In Prefect, you can add tasks to be run inside a flow. Tasks work similarly to flows since they are also defined with a decorator:


from prefect import flow
from prefect import task

def greet(name):
    print(f"Hello {name}")

def good_morning():

if __name__ == '__main__':

A flow can call another flow, and you can have as many layers as you want of flows calling flows inside them.

So you could change the code above to:

from prefect import flow

@flow # Now this is a flow
def greet(name):
    print(f"Hello {name}")

def good_morning():

if __name__ == '__main__':

A task cannot call another task inside itself.

1.2. Using tags

Adding tags in Prefect is quite straightforward:

from prefect import tags

with tags("env:pro", "type:greet"):
    good_morning() # Call any 'flow' you want

The tags are simply strings.

In order to categorize, I like to use tags with {key}:{value} format, like env:pro. This helps me to later filter flows.

1.3. Adding logs

To add logs linked to Prefect you need to:

from prefect import flow
from prefect import get_run_logger

def greet(name):
    logger = get_run_logger()"Hello {name}")

if __name__ == '__main__':

get_run_logger() must be called inside a task or a flow, else it will fail.

Also, notice that in this example, I defined a name for the flow.

2. Connection to Prefect Cloud

With the example we did Prefect run locally with a local database it created. This is useful for testing but you shouldn’t be running jobs in production without a server.

To connect to Prefect cloud you can do it with:

prefect cloud login

This will open an internet window where you will be asked to log in. Since this won’t work well on automated processes, the alternative is to create a Prefect token:

prefect cloud login -k your_token

Make sure to replace your_token. Prefect tokens usually start with pnu_.

In some cases, you might run into problems because Prefect is still using the default settings. To clean them out, go to ~/.prefect and delete the files that are there. Important: this will delete the local database with the information of the flows you just ran.

3. Using Prefect Blocks

You can use Domain LogoPrefect Blocks to set up connections to external tools. For example, you can use the Slack Block to store a Slack webhook. Once done, you can send Slack messages by simply running:

from prefect.blocks.notifications import SlackWebhook

slack_webhook_block = SlackWebhook.load("slack") # The name of the block you just created
slack_webhook_block.notify("Hello from Prefect!")

4. Naming conventions

When calling flows or tasks, I like to follow a naming convention where each unit inherits a prefix from its parent. In the example we used before, it would be:

from prefect import flow

def greet(name):
    print(f"Hello {name}")

def good_morning():

if __name__ == '__main__':

This helps keep things organized and makes it easier to understand the relation between different flows and tasks.

For a more complex example, let’s take a look at my personal pipeline Domain LogoVtasks. There we can see how the main flow vtasks calls different flows like vtasks.backup or vtasks.expensor. They in turn have different tasks inside:

- vtasks
  β”œβ”€β”€ vtasks.backup
  β”‚   β”œβ”€β”€ vtasks.backup.backup_files
  β”‚   β”œβ”€β”€ vtasks.backup.clean_backups
  β”‚   └── vtasks.backup.copy
  β”œβ”€β”€ vtasks.expensor
  β”‚   β”œβ”€β”€
  β”‚   └──
  └── ...

5. Migrating to prefect

I think that migrating from another orchestrator to Prefect can be easily done in incremental steps:

  1. Track executions
  2. Set up a Prefect Server (this is optional, you can use Prefect cloud instead)
  3. Create deployments so that you can trigger executions from Prefect
  4. Schedule flows with Prefect

The idea behind those steps is to be able to test Prefect and see if it indeed suits your needs better compared to other solutions.

5.1. Track executions with prefect

This first step towards implementing Prefect is to add some flows in any Python code you have. The idea is to use Prefect only for observability purposes so that you have a centralized place from where you can:

  • See logs
  • Track flows/tasks outcomes
  • Have notifications for failures

This won’t interfere with any existing orchestrator.

So for example let’s imagine that you use Airflow for triggering EMR jobs. Here what you would do is to add Prefect decorators to the code that runs in EMR.

For example you could have the following flows:

- my_emr_job
  β”œβ”€β”€ my_emr_job.transform
  β”‚   β”œβ”€β”€ my_emr_job.transform.rename
  β”‚   └── my_emr_job.transform.cast
  └── my_emr_job.write

This is just a dummy example but it helps to illustrate how you could add better observability.

5.2. Set up a Prefect server

At this point, you will need to decide if you want to set up your Prefect server. If you want to do so you can read how in Domain LogoPrefect Server Setup: Configuration and Deployment.

You can always start creating your Prefect server and then migrate to Prefect cloud or the other way around.

5.3. Create deployments for running flows

In this step, you will create Deployments for the flows you want to be able to trigger from Prefect.

This will allow you to perform backfills or any other manual run using Prefect UI.

Again, this won’t interfere with any existing orchestrator. It will only provide another way of manually triggering flows.

As an example, see how you can use dropdowns, date pickers, default values and other nice features that simplify triggering manual runs:

Create Prefect Work Pool

This is automatically created as long as you properly define the parameters in your flow. This is the code for the example above:

from datetime import datetime
from typing import Literal

def sf_export(
    env: Literal["pre", "snd", "pro"],
    select: str = None,
    exclude: str = None,
    min_date: datetime = None,
    max_date: datetime = None,
    timeout_mins: int = 150,
    version: str = "latest",

In order to see how to create the Deployments themselves, read Domain LogoPrefect Server Setup: Configuration and Deployment | Deployments

5.4. Schedule flows with prefect

The last step is to finally use Prefect for scheduling the flows.

Now is the time to finally switch to Prefect and to deprecate any other orchestrator you might have.

To see how to schedule flows, go to Domain LogoPrefect Server Setup: Configuration and Deployment | Deployments

6. Using prefect.client

We can use prefect.client to interact with the Prefect API. This allows us to do very powerful things like:

  • Updating tags at runtime based on input parameters
  • Querying past flows and taking certain actions based on the outcome

All interactions with prefect.client are async. That means you will need to do one of the following:

  1. Defining async functions whenever interact with it
  2. Calling (or await if you are in a Jupyter Notebook)

6.1. Updating tags at runtime

In general tags are defined prior to a flow_run. However, it is possible to modify those tags at runtime. This is very useful to add tags based on the input parameters of a flow. Imagine that we have a flow that can run in pre/pro, it is really useful to tag the flow_run with the environment where it ran.

We can update the tags with the following code:

import asyncio

from prefect import get_run_logger
from prefect.client import get_client
from prefect.context import get_run_context

def update_tags(tags):
    logger = get_run_logger()
    if not tags:
        logger.warning("No tags passed to 'update_tags', nothing to do")
        return True"Adding {tags=} to current flow_run")

    # Read current flow
    flow_run = get_run_context().flow_run
    tags += flow_run.tags

    client = get_client(), tags=set(tags)))
    return True

This is meant to be called from a flow. It won’t work if called from a task.

And then we would use it with:

from prefect import flow

def greet(name):
    print(f"Hello {name}")

def good_morning(env):

if __name__ == '__main__':

6.2. Querying prefect for conditional flows

You can query any prefect entity with the prefect.client like:

  • flows
  • flow_runs
  • tasks

For example for querying flow_runs we would use something like:

from prefect.client import get_client
from prefect.orion.schemas import filters
from prefect.orion.schemas import sorting

async def read_flow_runs(flow_run_filter=None, sort=sorting.FlowRunSort.START_TIME_DESC):
    Asynchronously retrieves flow runs with query limits.

        offset (int): Offset for paginating the results.
        flow_run_filter (prefect.orion.schemas.filters.FlowRunFilter, optional): Filter for flow runs. Defaults to None.
        sort (prefect.orion.schemas.sorting.FlowRunSort, optional): Sorting options. Defaults to sorting.FlowRunSort.START_TIME_DESC.

        list: List of flow runs.
    client = get_client()
    return await client.read_flow_runs(flow_run_filter=flow_run_filter, sort=sort)

We can use the flow_run_filter to filter the flow_runs we want to get. It is a little bit complex since you need to use the filter from prefect.orion.schemas.

For example:

async def query_flow_runs(name_like, env, state_names, start_time_min)
    filter_params = {
        "name": filters.FlowRunFilterName(like_=name_like),
        "tags": filters.FlowRunFilterTags(all_=[f"env:{env}"]),
        "state": filters.FlowRunFilterState(
        "start_time": filters.FlowRunFilterStartTime(after_=start_time_min),

    flow_run_filter = filters.FlowRunFilter(**filter_params)
    return await read_flow_runs(flow_run_filter)

Finally with that we can use that to see if we have already run a flow with:

from datetime import datetime, timedelta

flow_runs =
       env="pro", - timedelta(days=1)

if flow_runs:

This is very useful when you have a pipeline that executes multiple times a day and you don’t want to run a flow/task each time but only once if it hasn’t been done in less than a day.