Back to Openclaw

NVIDIA

docs/providers/nvidia.md

2026.5.53.1 KB
Original Source

NVIDIA provides an OpenAI-compatible API at https://integrate.api.nvidia.com/v1 for open models for free. Authenticate with an API key from build.nvidia.com.

Getting started

<Steps> <Step title="Get your API key"> Create an API key at [build.nvidia.com](https://build.nvidia.com/settings/api-keys). </Step> <Step title="Export the key and run onboarding"> ```bash export NVIDIA_API_KEY="nvapi-..." openclaw onboard --auth-choice nvidia-api-key ``` </Step> <Step title="Set an NVIDIA model"> ```bash openclaw models set nvidia/nvidia/nemotron-3-super-120b-a12b ``` </Step> </Steps> <Warning> If you pass `--nvidia-api-key` instead of the env var, the value lands in shell history and `ps` output. Prefer the `NVIDIA_API_KEY` environment variable when possible. </Warning>

For non-interactive setup, you can also pass the key directly:

bash
openclaw onboard --auth-choice nvidia-api-key --nvidia-api-key "nvapi-..."

Config example

json5
{
  env: { NVIDIA_API_KEY: "nvapi-..." },
  models: {
    providers: {
      nvidia: {
        baseUrl: "https://integrate.api.nvidia.com/v1",
        api: "openai-completions",
      },
    },
  },
  agents: {
    defaults: {
      model: { primary: "nvidia/nvidia/nemotron-3-super-120b-a12b" },
    },
  },
}

Built-in catalog

Model refNameContextMax output
nvidia/nvidia/nemotron-3-super-120b-a12bNVIDIA Nemotron 3 Super 120B262,1448,192
nvidia/moonshotai/kimi-k2.5Kimi K2.5262,1448,192
nvidia/minimaxai/minimax-m2.5Minimax M2.5196,6088,192
nvidia/z-ai/glm5GLM 5202,7528,192

Advanced configuration

<AccordionGroup> <Accordion title="Auto-enable behavior"> The provider auto-enables when the `NVIDIA_API_KEY` environment variable is set. No explicit provider config is required beyond the key. </Accordion> <Accordion title="Catalog and pricing"> The bundled catalog is static. Costs default to `0` in source since NVIDIA currently offers free API access for the listed models. </Accordion> <Accordion title="OpenAI-compatible endpoint"> NVIDIA uses the standard `/v1` completions endpoint. Any OpenAI-compatible tooling should work out of the box with the NVIDIA base URL. </Accordion> </AccordionGroup> <Tip> NVIDIA models are currently free to use. Check [build.nvidia.com](https://build.nvidia.com/) for the latest availability and rate-limit details. </Tip> <CardGroup cols={2}> <Card title="Model selection" href="/concepts/model-providers" icon="layers"> Choosing providers, model refs, and failover behavior. </Card> <Card title="Configuration reference" href="/gateway/configuration-reference" icon="gear"> Full config reference for agents, models, and providers. </Card> </CardGroup>