mkdocs/docs/en/deployment/docker-basic.md
This guide shows how to run Prompt Optimizer with Docker and how to configure custom OpenAI-compatible models with runtime environment variables.
docker run -d -p 8081:80 \
--restart unless-stopped \
--name prompt-optimizer \
linshen/prompt-optimizer:latest
Open the app at http://localhost:8081.
Custom models support the following runtime variables:
VITE_CUSTOM_API_KEY_<suffix>=your-api-key
VITE_CUSTOM_API_BASE_URL_<suffix>=your-base-url
VITE_CUSTOM_API_MODEL_<suffix>=your-model-name
VITE_CUSTOM_API_PARAMS_<suffix>=json-object-string
Notes:
KEY, BASE_URL, and MODEL are requiredPARAMS is optional and must be a JSON object stringPARAMS is injected into the final OpenAI-compatible request bodymodel, messages, and stream are ignored automaticallydocker run -d -p 8081:80 \
-e VITE_CUSTOM_API_KEY_nvidia=nvapi-xxx \
-e VITE_CUSTOM_API_BASE_URL_nvidia=https://integrate.api.nvidia.com/v1 \
-e VITE_CUSTOM_API_MODEL_nvidia=qwen/qwen3.5-397b-a17b \
-e 'VITE_CUSTOM_API_PARAMS_nvidia={"chat_template_kwargs":{"enable_thinking":true},"temperature":0.6,"top_p":0.95,"max_tokens":16384}' \
--restart unless-stopped \
--name prompt-optimizer \
linshen/prompt-optimizer:latest
This is useful for:
temperature, top_p, and max_tokenschat_template_kwargsservices:
prompt-optimizer:
image: linshen/prompt-optimizer:latest
ports:
- "8081:80"
restart: unless-stopped
environment:
VITE_CUSTOM_API_KEY_nvidia: nvapi-xxx
VITE_CUSTOM_API_BASE_URL_nvidia: https://integrate.api.nvidia.com/v1
VITE_CUSTOM_API_MODEL_nvidia: qwen/qwen3.5-397b-a17b
VITE_CUSTOM_API_PARAMS_nvidia: '{"chat_template_kwargs":{"enable_thinking":true},"temperature":0.6,"top_p":0.95,"max_tokens":16384}'
PARAMS is not valid JSON, the model still loads, but the extra parameters are ignored.