Back to Tensorzero

TensorZero Recipe: DPO (Preference Fine-tuning) with OpenAI

recipes/dpo/openai/README.md

2026.4.1531 B
Original Source

TensorZero Recipe: DPO (Preference Fine-tuning) with OpenAI

The openai.ipynb notebook provides a step-by-step recipe to perform Direct Preference Optimization (DPO) — also known as Preference Fine-tuning — of OpenAI models based on data collected by the TensorZero Gateway.

Set TENSORZERO_CLICKHOUSE_URL=http://chuser:chpassword@localhost:8123/tensorzero and OPENAI_API_KEY in the shell your notebook will run in.

Setup

We recommend using uv.

bash
uv sync