mkdocs/docs/en/deployment/docker-advanced.md
This page covers the advanced runtime configuration pattern for custom OpenAI-compatible models in Docker deployments.
When a provider needs more than apiKey, baseURL, and model, use:
VITE_CUSTOM_API_PARAMS_<suffix>=json-object-string
Example:
services:
prompt-optimizer:
image: linshen/prompt-optimizer:latest
environment:
VITE_CUSTOM_API_KEY_nvidia: nvapi-xxx
VITE_CUSTOM_API_BASE_URL_nvidia: https://integrate.api.nvidia.com/v1
VITE_CUSTOM_API_MODEL_nvidia: qwen/qwen3.5-397b-a17b
VITE_CUSTOM_API_PARAMS_nvidia: '{"chat_template_kwargs":{"enable_thinking":true},"temperature":0.6,"top_p":0.95,"max_tokens":16384}'
temperature, top_p, and max_tokenschat_template_kwargsPARAMS must be a JSON object stringmodel, messages, and stream are ignored automaticallytimeout is allowed and can be used to override request timeout