docs/current_docs/reference/configuration/custom-runner.mdx
A runner is the "backend" of Dagger where containers are actually executed.
Runners are responsible for:
The runner is distributed as a container image, making it easy to run on various container runtimes like Docker, Kubernetes, Podman, etc.
The consolidated steps to use a custom runner are:
v0.14.0 Python SDK, you should use
v0.14.0 of the engine._EXPERIMENTAL_DAGGER_RUNNER_HOST environment variable with a value pointing to your target.dagger call or execute SDK code directly with that environment variable set.:::important
The _EXPERIMENTAL_DAGGER_RUNNER_HOST variable is experimental and may change in future.
:::
The runner is distributed as a container image at registry.dagger.io/engine.
v0.12.3 release has a corresponding image at registry.dagger.io/engine:v0.12.3CAP_SYS_ADMIN, in order to execute workflows. For example, this will be granted when using the --privileged flag of docker run./var/lib/dagger.
/var/lib/dagger is itself an overlayfs.docker run command as -v dagger-engine:/var/lib/dagger.To configure a manually started Dagger Engine, see the Dagger Engine configuration documentation.
After the runner starts up, the CLI needs to connect to it. In the default situation, this will happen automatically.
However, if the _EXPERIMENTAL_DAGGER_RUNNER_HOST environment variable is set,
then the CLI will instead connect to the endpoint specified there. This
environment variable currently accepts values in the following format:
container://<container name> - Connect to the runner inside the given host container.
container+<runtime>://<container name>, e.g. container+podman://dagger-engine.image://<container image reference> - Start the runner in Docker using the provided container image, pulling it locally if needed
image+<runtime>://<container image reference>, e.g. image+podman://registry.dagger.io/engine:latest.kube-pod://<podname>?context=<context>&namespace=<namespace>&container=<container> - Connect to the runner inside the given Kubernetes pod.
unix://<path to unix socket> - Connect to the runner over the provided UNIX socket.tcp://<address:port> - Connect to the runner over TCP using the provided address and port.:::warning Dagger itself does not set up any encryption of data sent "over the wire". It relies on the underlying connection type to implement this when needed. If you are using a connection type that does not provide encryption, then all queries and responses will be sent in plaintext over the wire from the Dagger CLI to the runner. :::
:::warning GPU support is currently experimental and only works with NVIDIA GPUs. :::
In order to use a GPU, Dagger needs a custom, GPU-enabled runner and the NVIDIA Container Toolkit.
Assuming that Dagger and the NVIDIA Container Toolkit are already installed on a GPU-capable host, use the instructions below to replace the default runner with a GPU-enabled runner.
:::note The sections below provide instructions for the local host and for cloud infrastructure providers Fly.io and Lambda Labs. These instructions can be adapted for use on other cloud providers, so long as the host has access to an NVIDIA GPU. :::
<Tabs> <TabItem value="Local host">Use the following commands to deploy a GPU-enabled Dagger runner on the local host:
VERSION=$(dagger version | cut -d' ' -f2)
docker rm -f dagger-engine-${VERSION} 2>/dev/null && docker run --gpus all -d --privileged -e _EXPERIMENTAL_DAGGER_GPU_SUPPORT=true --name dagger-engine-${VERSION} registry.dagger.io/engine:${VERSION}-gpu -- --debug
Use the following commands to deploy a GPU-enabled Dagger runner on Fly.io:
export FLYIO_TOKEN=YOUR-FLY.IO-TOKEN
export FLYIO_ORG=YOUR-FLY.IO-ORG-NAME
dagger -m github.com/samalba/dagger-modules/nvidia-gpu call deploy-dagger-on-fly --token env://FLYIO_TOKEN --org env://FLYIO_ORG
:::note
By default, the previous Dagger Function call uses the region ord and an NVIDIA L40s GPU, but you can easily fork and modify the module code for other requirements.
:::
The Dagger Function returns a message in your terminal, with an environment variable to export. For example:
export _EXPERIMENTAL_DAGGER_RUNNER_HOST=tcp://dagger-v0-14-0-smart-gerhard-2024-12-06.internal:2345
Copy and paste the previous command in your active terminal. From now on, Dagger will execute all function calls using the remote Dagger Engine running on Fly.io.
To stop using the remote Dagger Engine on Fly.io and return to using your local Dagger Engine, use the following commands:
unset _EXPERIMENTAL_DAGGER_RUNNER_HOST
# Make sure the Fly app name matches the one that was provisioned earlier
dagger -m github.com/samalba/dagger-modules/nvidia-gpu call destroy-dagger-on-fly --token env://FLYIO_TOKEN --app dagger-v0-14-0-smart-gerhard-2024-12-06
Use the following commands to deploy a GPU-enabled Dagger runner on a Lambda Labs virtual machine:
VERSION=$(dagger version | cut -d' ' -f2)
docker rm -f dagger-engine-${VERSION} 2>/dev/null && docker run --gpus all -d --privileged -e _EXPERIMENTAL_DAGGER_GPU_SUPPORT=true --name dagger-engine-${VERSION} registry.dagger.io/engine:${VERSION}-gpu -- --debug
:::note
The default user ubuntu does not have access to the Docker socket. Fix this with the command sudo usermod -aG docker ubuntu, restart the shell session, and try the command again.
:::
Once your GPU-enabled Dagger runner is configured, use the following Dagger Function to test if Dagger has access to the GPU:
dagger -m github.com/samalba/dagger-modules/nvidia-gpu call has-gpu
If the GPU is properly configured, this Dagger Function returns true.
Here's a more complex Dagger Function:
dagger -m github.com/samalba/dagger-modules/nvidia-gpu call ollama-run --prompt "What color is the sky?"
This Dagger Function sets up an Ollama server, pulls a model, and prompts it with a question. It returns the response query from the prompt passed as argument.