1. Resources
  2. /
  3. Blog
  4. /
  5. Fully dynamic pipelines with Bazel and Buildkite

Fully dynamic pipelines with Bazel and Buildkite

15 minute read

Every Bazel project begins with one little command:

1
bazel build //...

If you've used Bazel before, you know that's the command that tells Bazel to build everything in the workspace. It's convenient, and it works: depending on how many targets you're building (and how well you're caching them), you may be able to get by with that one command for a good while.

But at a certain scale, building the whole Bazel workspace may no longer make sense. It may be when your builds become so complex that they exhaust all available resources, and things start to slow down—or just fall over. Or it may be when you decide to build more efficiency into your process, building only the targets that need to be built from one commit to the next.

Bazel itself offers a ton of flexibility for running highly-focused, selective builds, from its precise target patterns to powerful tools like bazel query. But to do selective builds well, you'll often need more than Bazel alone—you'll need deep flexibility at the CI layer as well. Building an adaptable pipeline that can take full advantage of Bazel at scale is a tough problem to solve—and it's even tougher when the underlying platform requires that you define the behavior of that pipeline statically, and up front, with YAML and Bash.

In this hands-on post, you'll see how with Bazel and Buildkite, you can approach this problem differently. Specifically, you'll learn how to:

  • Combine Git with bazel query to identify which Bazel targets were changed in a given commit
  • Use Python to define a fully dynamic, adaptable pipeline that builds only the Bazel packages that need to be built, adding additional steps to the pipeline at runtime as needed
  • Capture the details of each Bazel build, transforming Bazel's raw build events into rich annotations that improve visibility and tighten feedback loops

We've got a lot to cover—so get ready to download some tools, edit some code, and run some commands that'll have you driving your pipelines dynamically with Bazel in no time.

Let's get started.

Installing prerequisites

If you plan to work through this example (and I hope you will!), you'll need to set up a few things first:

  1. Bazel: We recommend installing Bazel with Bazelisk. If you're on a Mac and have Homebrew installed, you can do that by running brew install bazelisk. If not, follow the instructions for your operating system.
  2. Python: Any recent version should do. You'll need Python to run the code that generates the Buildkite pipeline definition.
  3. The Buildkite agent: The agent is a lightweight binary that connects to Buildkite to run your builds. Since we'll be running the builds for this walkthrough on your local machine, you'll need to install the agent so you can run it later on. If you're on a Mac, you can do that by running brew install buildkite/buildkite/buildkite-agent.

Make sure everything's set up correctly before moving on:

1
2
3
4
5
6
7
8
bazel --version
bazel 7.4.1

buildkite-agent --version                                      
buildkite-agent version 3.95.1

python3 --version
Python 3.13.1

Getting the code

Rather than create everything from scratch, we'll use an existing repository to get your project properly bootstrapped so you can follow along easily. You'll find that repository on GitHub:

https://github.com/cnunciato/bazel-buildkite-example

You'll also be triggering pipeline builds based on GitHub commits, so you'll need to get a copy of the example repository into your GitHub account as well, and then clone your remote copy so you can push to it directly.

The easiest way to do that is to create a new repository from the example template, then clone it to your local machine in the usual way. If you happen to have the GitHub CLI installed, you can do that with a single command:

1
2
3
4
gh repo create bazel-buildkite-example \
  --template cnunciato/bazel-buildkite-example \
  --public \
  --clone

With your copy of the repository created and cloned locally, change to it to get started:

1
2
3
4
cd bazel-buildkite-example

ls
.buildkite app  library  MODULE.bazel  README.md

Let's have a look at the contents of the repository next.

Understanding the repository structure

The example we're working with is a simple Python monorepo that contains two Bazel packages:

  • a Python library package named library
  • a Python "binary" package (really just a Python script) named app that depends on the library package

Each has its own BUILD.bazel file of course, and a MODULE.bazel file defines the surrounding Bazel workspace. (We'll get to the .buildkite folder in a moment.) Here's the full tree:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
bazel-buildkite-example/
├── .buildkite               # Buildkite configuration
│   ├── pipeline.yml     
│   ├── pipeline.py        
│   ├── step.py     
│   └── utils.py           
├── app                       # Python application that depends on the library
│   ├── BUILD.bazel.       
│   ├── main.py
│   └── test_main.py
└── library                   # Python library  
    ├── BUILD.bazel
    ├── hello.py
    └── test_hello.py
└── MODULE.bazel              # Bazel module definition

The application and the library

The library package exposes a single Python function whole sole responsibility is to deliver the greeting we all know so well:

1
2
3
# library/hello.py
def get_greeting():
    return "Hello, world!"

The app package imports and uses that library by calling get_greeting() and using the result to print a message to the terminal:

1
2
3
4
5
6
7
8
# app/main.py
from library.hello import get_greeting

def say_hello():
    response = get_greeting()
    return f"The Python library says: '{response}'"

print(say_hello())

That's about it for the application—again, it's intentionally simple. What's important is that it sets up the dependency relationship we'll be using to illustrate the example, which is explicitly defined in the app package's BUILD.bazel file:

1
2
3
4
5
6
7
8
9
10
# app/BUILD.bazel
load("@rules_python//python:defs.bzl", "py_binary", "py_test")

py_binary(
    name = "main",
    srcs = ["main.py"],
    deps = [
        "//library:hello", # 👈 This tells Bazel that `app` depends on `library`.
    ],
)

Our goal is to draw on the existence of this dependency to implement the logic that'll compute the pipeline dynamically from one commit to the next. To do that, we'll use Bazel (specifically bazel query, as you'll see in a moment) to figure out whether to add an additional step to the pipeline to build and test the app package whenever something changes in library.

Go ahead and confirm this all works by running the app package with Bazel now:

1
2
3
4
bazel run //app:main
...
INFO: Running command line: bazel-bin/app/main
The Python library says: 'Hello, world!'

You can confirm the dependency relationship as well by asking Bazel which other packages depend on any targets in //library:

1
2
bazel query "kind('py_binary', rdeps(//..., //library/...))"
//app:main

Now let's take a closer look at what's happening in the .buildkite folder.

The Buildkite pipeline definition

The core of the approach we're taking in this example is to use Bazel (in combination with Git) to assemble the Buildkite pipeline dynamically based on the Bazel dependency graph. There are three files that conspire to make that happen.

The entrypoint: pipeline.yml

This is the file that kicks off the process. When a build job starts, the Buildkite agent checks out your source code, finds this file, and evaluates it, running the commands listed in the first step and passing the results to buildkite-agent pipeline upload:

1
2
3
4
5
# .buildkite/pipeline.yml
steps:
  - label: ":python: Compute the pipeline with Python"
    commands:
      - python3 .buildkite/pipeline.py | buildkite-agent pipeline upload

This step runs the Python script that computes the work to be done in the pipeline run and passes the result as JSON to buildkite-agent, which uploads it to Buildkite, appending it to the already-running pipeline.

The pipeline generator: pipeline.py

This is the Python script that does the work of assembling the pipeline programmatically.

Here's how it works:

  1. It begins by using Git to identify the directories that changed in the most recent commit, then runs bazel query to identify which of those directories, if any, contain Bazel packages.
  2. For each changed package, it adds a step to the pipeline to build and test all of the Bazel targets in the package.
  3. If any of those packages were Python libraries with one or more dependents, it adds a command to be run after that package's bazel build to generate and append a follow-up step to the pipeline to build each of the library's dependents as well.
  4. Writes the resulting pipeline as a JSON string to stdout.

Open pipeline.py in your favorite editor for a closer look. The inline comments there (and below) should clarify what's happening in more detail:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
# .buildkite/pipeline.py
from utils import run, filter_dirs, get_paths, get_package_step, to_json

# By default, do nothing.
steps = []

# Get a list of directories changed in the most recent commit.
changed_paths = run(["git", "diff-tree", "--name-only", "HEAD~1..HEAD"])
changed_dirs = filter_dirs(changed_paths)

# Query the Bazel workspace for a list of all packages (libraries, binaries, etc.).
all_packages = run(["bazel", "query", "'/...'"])

# Using both lists, figure out which packages need to be built.
changed_packages = [p for p in changed_dirs if p in get_paths(all_packages)]

# For each changed Bazel package, assemble a pipeline step programmatically to
# build and test all of its targets. For Python libraries, add a follow-up step
# to be run later that builds and tests their reverse dependencies as well.
for pkg in changed_packages:

    # Make a step that runs `bazel build` and `bazel test` for this package.
    package_step = get_package_step(pkg)

    # Use Bazel to query the package for any Python libraries.
    libraries = run(["bazel", "query", f"kind(py_library, '//{pkg}/...')"])

    for lib in libraries:

        # Find the library's reverse dependencies.
        reverse_deps = run(["bazel", "query", f"rdeps(//..., //{pkg}/...)"])

        # Filter this list to exclude any package that's already set to be built. 
        reverse_deps_to_build = [
            p for p in get_paths(reverse_deps, pkg) if p not in changed_packages
        ]

        for dep in reverse_deps_to_build:
            rdep_step = get_package_step(dep)

            # Add a command to the library's command list to generate and append
            # a build step (at runtime) for the dependent package as well.
            package_step["commands"].extend([
                f"echo 'Generating and uploading a follow-up step to build {dep}...'",
                f"python3 .buildkite/step.py {dep} | buildkite-agent pipeline upload"
            ])

    # Add this package step to the pipeline.
    steps.append(package_step)

# Emit the pipeline as JSON to be uploaded to Buildkite.
print(to_json({"steps": steps}, 4))

You should be able to run this script now to see that it works:

1
2
3
4
python3 .buildkite/pipeline.py  
{
    "steps": []
}

If you're wondering why the steps array is empty, it's because on my machine, the latest commit in the repository touches neither app nor library:

1
2
git show --name-only
README.md

Which is just what we want—namely to avoid wasting time and compute resources doing work we don't have to. Here, since there isn't anything Bazel-buildable that's changed, there's no need to run bazel at all, so the resulting pipeline is empty. Later, though, when there is, the logic we've written will recognize that and do what's expected, but nothing more.

Utility functions: utils.py

This file just defines a few helper functions for pipeline.py, most of which handle common tasks like running shell commands (git, bazel), processing lists, serializing JSON, and the like, all in the interest of making pipeline.py more readable and maintainable.

Two of those functions, however, are worth calling out:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
# .buildkite/utils.py
import json, os, subprocess

# Returns a Buildkite `command` step as a Python dictionary.
def command_step(emoji, label, commands=[], plugins=[]):
    step = {"label": f":{emoji}: {label}", "commands": commands}
    if plugins:  
        step["plugins"] = plugins
    return step
    
# Returns a Buildkite `command` step that builds, tests, and annotates a Bazel package.
def get_package_step(package):
    return command_step(
        "bazel",
        f"Build and test //{package}/...",
        [
            f"bazel test //{package}/...",
            f"bazel build //{package}/... --build_event_json_file=bazel-events.json",
        ],
        [{ "bazel-annotate#v0.1.0": { "bep_file": f"bazel-events.json"} }],
    )

These two functions—given a package name, label, and emoji (always!)—produce the individual steps that ultimately make up the full pipeline definition:

  • command_step() returns a Buildkite step-shaped Python dictionary to be converted to JSON later with pipeline.py.
  • get_package_step() calls command_step() to assemble a step that runs bazel test and bazel build for the specified package. The build step also tells Bazel to produce a build-event file (or BEP file) containing the details of the build, which we'll use (via the official bazel-annotate plugin) to render a rich annotation of the build in the Buildkite dashboard.

You'll see how this all comes together in the next section.

Incidentally, why Python?

We chose Python for this walkthrough because it's well known and easy to read. The mechanics are the same for any language, though; as long as your language of choice can produce JSON or YAML, you can use it to generate pipelines in Buildkite. See the Dynamic Pipelines docs and the Buildkite SDK for examples in other languages.

Running it locally: Dynamic pipelines in action

Let's run through a couple of scenarios locally to get a sense of how the pipeline will react to different types of changes.

Scenario 1: Changes to the application only

First, simulate a change to the application by adding a comment to ./app/main.py, committing, and then running the pipeline generator:

1
2
3
4
5
echo "# Adding a comment" >> app/main.py
git add app/main.py
git commit -m "Update the app"

python3 .buildkite/pipeline.py

The output should look something like this:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
{
    "steps": [
        {
            "label": ":bazel: Build and test //app/...",
            "commands": [
                "bazel test //app/...",
                "bazel build //app/... --build_event_json_file=bazel-events.json"
            ],
            "plugins": [
                {
                    "bazel-annotate#v0.1.0": {
                        "bep_file": "bazel-events.json"
                    }
                }
            ]
        }
    ]
}

Notice the pipeline reflects that only the app package will be built and tested: the script correctly detected that no other packages were changed.

Scenario 2: Changes to the shared library

Now try making a change to the library package:

1
2
3
4
5
echo "# Adding a comment" >> library/hello.py
git add library/hello.py
git commit -m "Update the library"

python3 .buildkite/pipeline.py

This time, the output should include steps for both the library and the app—the latter as a follow-up step once the library build finishes (the JSON for that step passed as input to buildkite-agent pipeline upload):

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
{
    "steps": [
        {
            "label": ":bazel: Build and test //library/...",
            "commands": [
                "bazel test //library/...",
                "bazel build //library/... --build_event_json_file=bazel-events.json",
                
                "echo '👇 Generating and uploading a follow-up step to build app...'",
                "python3 .buildkite/step.py app | buildkite-agent pipeline upload"
            ],
            "plugins": [
                {
                    "bazel-annotate#v0.1.0": {
                        "bep_file": "bazel-events.json"
                    }
                }
            ]
        }
    ]
}

Using Git, Bazel, and a few Bazel queries, the pipeline generator correctly detected that because the library package had changed, both app and library should be built and tested, and that app should be built as a follow-up step.

Bringing it all together

With things looking good locally, it's time to push some commits to GitHub and run some real builds.

Get the Buildkite agent started

You'll be running the Buildkite agent locally, and for that, you'll need to give it a token that tells it which cluster and build-event queue to subscribe to. Here's how to do that:

  1. If you don't yet have a Buildkite account, sign up for a free trial. Give your organization a name, pick the Pipelines path, choose Create pipeline > Create starter pipeline > Set up local agent, and follow the instructions to start buildkite-agent.
  2. If you do have a Buildkite account, navigate to Agents in the Buildkite dashboard, choose (or create) a self-hosted cluster, then choose Agent tokens > New token. Copy the generated token to your clipboard, then return to your terminal to start buildkite-agent.
1
buildkite-agent start --token ${your-token}

At this point, you should have a Buildkite agent running locally:

The Buildkite agent running locally

The Buildkite agent running locally

Create a new pipeline

  1. In the Buildkite dashboard, navigate to Pipelines, then New pipeline. Connect the pipeline to your GitHub account if you're prompted to do so, making sure to grant access to your newly created bazel-buildkite-example repository. (This is important.)
  2. On the New Pipeline page, connect your repository, choose HTTPS for the checkout type, name the pipeline bazel-buildkite-example, and choose the Cluster you selected (or that was created for you) in the previous section. Leave the default pipeline steps as they are, then choose Create pipeline.

With the pipeline created, and the buildkite-agent connected and listening in your terminal, it's time to push some commits to see how this works end to end.

Run some builds 🚀

Assuming you've been following along step by step, your most recent local commit should still be the one you made to the library package above—which is good, because that's exactly the one we want to use to validate the logic we care about. Open another terminal tab to confirm that:

1
2
3
4
5
git log -1

commit 91929ffd46cb530669904b42e8da40c512f5be02 (HEAD -> main)
...
Update the library 

Go ahead and push that commit to GitHub now (straight to main, for simplicity) to trigger a new build of the bazel-buildkite-example pipeline:

1
git push origin main

You should see your locally running Buildkite agent respond immediately to pick up the job and begin processing it.

Now, back in the Buildkite dashboard, navigate to the pipeline and watch the build as it unfolds, one dynamically added step at a time, concluding with the app package being built and tested as a downstream dependent as expected:

A multi-step dynamic pipeline computed from Bazel package dependencies

Feel free to experiment here, making additional commits and pushing them as often as you like to see how the pipeline responds. In general, you should see that:

  • Commits that change both the app and library packages trigger pipelines that build both packages
  • Commits that change only the app package trigger pipelines that build only that package
  • Commits that change only the library package build the library first, then the app—again, as a downstream dependent, conforming to the business logic we set out to implement
  • Commits to anything else are ignored, and the pipeline completes within a few seconds

And there you have it: a fully dynamic, easily maintainable and extensible—and testable!—pipeline that uses Bazel, Git, and Python to get the job done—no YAML required.

Capture and convert Bazel events into rich annotations

Before we wrap up, let's come back to those build options we've been passing into our bazel build commands:

1
bazel build //app/... --build_event_json_file=bazel-events.json

That --build_event_json_file option tells Bazel to collect and emit structured data about the build process, such as which targets were affected, their types, how long it took to build each one, and whether the build passed or failed. These build events are protocol-buffer messages that Bazel writes to a text file containing one line per event, each individual line a variably structured JSON object.

Here's a snippet from one of my own BEP files, for example:

1
2
3
4
5
{"id":{"started":{}},"children":[{"progress":{}},{"unstructuredCommandLine":{}},{"structuredCommandLine":{"commandLineLabel":"original"}},{"structuredCommandLine":{"commandLineLabel":"canonical"}},{"structuredCommandLine":{"commandLineLabel":"tool"}},{"buildMetadata":{}},{"optionsParsed":{}},{"workspaceStatus":{}},{"pattern":{"pattern":["//..."]}},{"buildFinished":{}}],"started":{"uuid":"d7ak38s7-d3b2-45ae-a79c-63eks82b2c82","startTimeMillis":"1743442193828","buildToolVersion":"7.4.1","optionsDescription":"--test_output\u003dALL --build_event_json_file\u003dbazel-events.json","command":"build","workingDirectory":"/Users/cnunciato/Projects/cnunciato/bazel-buildkite-example-fromtemplate","workspaceDirectory":"/Users/cnunciato/Projects/cnunciato/bazel-buildkite-example-fromtemplate","serverPid":"20646","startTime":"2025-03-31T17:29:53.828Z"}}
{"id":{"buildMetadata":{}},"buildMetadata":{}}
{"id":{"structuredCommandLine":{"commandLineLabel":"tool"}},"structuredCommandLine":{}}
{"id":{"pattern":{"pattern":["//..."]}},"children":[{"targetConfigured":{"label":"//app:main"}},{"targetConfigured":{"label":"//app:test_main"}},{"targetConfigured":{"label":"//library:hello"}},{"targetConfigured":{"label":"//library:hello_wheel"}},{"targetConfigured":{"label":"//library:hello_wheel_dist"}},{"targetConfigured":{"label":"//library:test_hello"}}],"expanded":{}}
{"id":{"progress":{}},"children":[{"progress":{"opaqueCount":1}},{"workspace":{}}],"progress":{"stderr":"\u001b[32mComputing main repo mapping:\u001b[0m \n\r\u001b[1A\u001b[K\u001b[32mLoading:\u001b[0m \n\r\u001b[1A\u001b[K\u001b[32mLoading:\u001b[0m 0 packages loaded\n"}}

These events can be incredibly helpful for understanding what happened during a given build—but in order to use them, you need to parse them, store them somewhere, and somehow transform them into something readable that your team can review on a regular basis.

Buildkite annotations are a great way to make use of this data. Annotations are essentially Markdown snippets that you can attach to a pipeline build with buildkite-agent annotate:

1
buildkite-agent annotate ":bazel: Hello from Bazel 👋"

You can build out your own JSON parsing and annotation logic if you like—see the Bazel team's own Buildkite pipeline for an example. Or you can do what we've done here, and just use the official Bazel BEP Annotate plugin, which understands the BEP file format and handles everything for you:

A Buildkite annotation showing the results of a Bazel build, sourced from a Bazel event protocol (BEP) file.

A Buildkite annotation showing the results of a Bazel build, sourced from a Bazel event protocol (BEP) file.

See the annotation API docs for more details.

Next steps

We covered a lot in this post—and hopefully you've now got a good sense of what's possible when you move away from writing your pipelines in static languages like YAML and toward driving them dynamically with tools like Bazel and Buildkite.

To keep the learning going:


Related posts

Start turning complexity into an advantage

Create an account to get started with a 30-day free trial. No credit card required.

Buildkite Pipelines

Platform

  1. Pipelines
  2. Pipeline templates
  3. Public pipelines
  4. Test Engine
  5. Package Registries
  6. Mobile Delivery Cloud
  7. Pricing

Hosting options

  1. Self-hosted agents
  2. Mac hosted agents
  3. Linux hosted agents

Resources

  1. Docs
  2. Blog
  3. Changelog
  4. Webinars
  5. Plugins
  6. Case studies
  7. Events
  8. Comparisons

Company

  1. About
  2. Careers
  3. Press
  4. Brand assets
  5. Contact

Solutions

  1. Replace Jenkins
  2. Workflows for AI/ML
  3. Testing at scale
  4. Monorepo mojo
  5. Bazel orchestration

Support

  1. System status
  2. Forum
© Buildkite Pty Ltd 2025