crates/tensorzero-client/examples/inference_demo/README.md
This demo shows performing inference with both an HTTP gateway server and an embedded gateway
docker compose up in <tensorzero_repository>/examples/haiku-hidden-preferencesThe following steps should be run from the root of the repository
cargo run --example inference_demo -- --gateway-url http://localhost:3000 --function-name 'judge_haiku' --streaming '{"topic": "Rivers", "haiku": "Endless roaring flow. Mountains weep streams for oceans. Carve earth like giants"}'
CLICKHOUSE_URL=http://127.0.0.1:8123/tensorzero cargo run --example inference_demo -- --config-path examples/haiku-hidden-preferences/config/tensorzero.toml --function-name judge_haiku --streaming '{"topic": "Rivers", "haiku": "Endless roaring flow. Mountains weep streams for oceans. Carve earth like giants"}'
The '--streaming' flag controls whether or not the output is streamed to the console as it becomes available, or only disabled when the full response is available.