docs/customize/model-providers/more/ovhcloud.mdx
OVHcloud AI Endpoints is a serverless inference API that provides access to a curated selection of models (e.g., Llama, Mistral, Qwen, Deepseek). It is designed with security and data privacy in mind and is compliant with GDPR.
<Info> To get started, create an API key on the OVHcloud [AI Endpoints website](https://endpoints.ai.cloud.ovh.net/). For more information, including pricing, visit the OVHcloud [AI Endpoints product page](https://www.ovhcloud.com/en/public-cloud/ai-endpoints/). </Info>We recommend configuring Qwen2.5-Coder-32B-Instruct as your chat model. Check our catalog to see all of our models hosted on AI Endpoints.
OVHcloud AI Endpoints provides access to the following models:
Llama Models:
llama3.1-8b - Llama 3.1 8B Instruct (supports function calling)llama3.1-70b - Llama 3.1 70B Instructllama3.3-70b - Llama 3.3 70B Instruct (supports function calling)Qwen Models:
qwen2.5-coder-32b - Qwen 2.5 Coder 32B Instruct (supports function calling)qwen3-32b - Qwen 3 32B (supports function calling)qwen3-coder-30b-a3b - Qwen 3 Coder 30B A3B Instruct (supports function calling)qwen2.5-vl-72b - Qwen 2.5 VL 72B Instruct (vision-language model)Mistral Models:
mistral-7b - Mistral 7B Instruct v0.3mistral-8x7b - Mixtral 8x7B Instruct v0.1mistral-nemo - Mistral Nemo Instruct 2407 (supports function calling)mistral-small-3.2-24b - Mistral Small 3.2 24B Instruct (supports function calling)OpenAI Models:
gpt-oss-20b - GPT-OSS 20B (supports function calling)gpt-oss-120b - GPT-OSS 120B (supports function calling)DeepSeek Models:
DeepSeek-R1-Distill-Llama-70B - DeepSeek R1 Distill Llama 70B (supports function calling)Other Models:
codestral-mamba-latest - Codestral Mamba 7B v0.1Many OVHcloud models support function calling (tool use), which enables the model to interact with external tools and APIs. Models with function calling support include:
Function calling is automatically enabled for supported models when using Continue.
<Tabs> <Tab title="YAML"> ```yaml title="config.yaml" name: My Config version: 0.0.1 schema: v1models:
- name: Qwen2.5-Coder-32B-Instruct
provider: ovhcloud
model: qwen2.5-coder-32b
apiKey: <YOUR_AIENDPOINTS_API_KEY>
```
</Tab>
<Tab title="JSON">
```json title="config.json"
{
"models": [
{
"title": "Qwen2.5-Coder-32B-Instruct",
"provider": "ovhcloud",
"model": "qwen2.5-coder-32b",
"apiKey": "<YOUR_AIENDPOINTS_API_KEY>"
}
]
}
```
</Tab>
Here's an example configuration for a model that supports function calling:
<Tabs> <Tab title="YAML"> ```yaml title="config.yaml" name: My Config version: 0.0.1 schema: v1models:
- name: GPT-OSS-120B
provider: ovhcloud
model: gpt-oss-120b
apiKey: <YOUR_AIENDPOINTS_API_KEY>
capabilities:
- tool_use
```
</Tab>
<Tab title="JSON">
```json title="config.json"
{
"models": [
{
"title": "GPT-OSS-120B",
"provider": "ovhcloud",
"model": "gpt-oss-120b",
"apiKey": "<YOUR_AIENDPOINTS_API_KEY>",
"capabilities": ["tool_use"]
}
]
}
```
</Tab>
We recommend configuring bge-multilingual-gemma2 as your embeddings model.
<Tabs> <Tab title="YAML"> ```yaml title="config.yaml" name: My Config version: 0.0.1 schema: v1models: - name: BGE Multilingual Gemma2 provider: ovhcloud model: bge-multilingual-gemma2 apiKey: <YOUR_AIENDPOINTS_API_KEY> roles: - embed
</Tab>
<Tab title="JSON">
```json title="config.json"
{
"embeddingsProvider": {
"provider": "ovhcloud",
"model": "bge-multilingual-gemma2",
"apiKey": "<YOUR_AIENDPOINTS_API_KEY>"
}
}