| title | OVHcloud |
|---|---|
| description | Configure OVHcloud AI Endpoints with Continue to access their GDPR-compliant serverless inference API for models like Qwen, Llama, Mistral, and Deepseek, with strong security and data privacy features |
OVHcloud AI Endpoints is a serverless inference API that provides access to a curated selection of models (e.g., Llama, Mistral, Qwen, Deepseek). It is designed with security and data privacy in mind and is compliant with GDPR.
To get started, create an API key on the OVHcloud [AI Endpoints website](https://endpoints.ai.cloud.ovh.net/). For more information, including pricing, visit the OVHcloud [AI Endpoints product page](https://www.ovhcloud.com/en/public-cloud/ai-endpoints/).We recommend configuring Qwen2.5-Coder-32B-Instruct as your chat model. Check our catalog to see all of our models hosted on AI Endpoints.
OVHcloud AI Endpoints provides access to the following models:
Llama Models:
llama3.1-8b- Llama 3.1 8B Instruct (supports function calling)llama3.1-70b- Llama 3.1 70B Instructllama3.3-70b- Llama 3.3 70B Instruct (supports function calling)
Qwen Models:
qwen2.5-coder-32b- Qwen 2.5 Coder 32B Instruct (supports function calling)qwen3-32b- Qwen 3 32B (supports function calling)qwen3-coder-30b-a3b- Qwen 3 Coder 30B A3B Instruct (supports function calling)qwen2.5-vl-72b- Qwen 2.5 VL 72B Instruct (vision-language model)
Mistral Models:
mistral-7b- Mistral 7B Instruct v0.3mistral-8x7b- Mixtral 8x7B Instruct v0.1mistral-nemo- Mistral Nemo Instruct 2407 (supports function calling)mistral-small-3.2-24b- Mistral Small 3.2 24B Instruct (supports function calling)
OpenAI Models:
gpt-oss-20b- GPT-OSS 20B (supports function calling)gpt-oss-120b- GPT-OSS 120B (supports function calling)
DeepSeek Models:
DeepSeek-R1-Distill-Llama-70B- DeepSeek R1 Distill Llama 70B (supports function calling)
Other Models:
codestral-mamba-latest- Codestral Mamba 7B v0.1
Many OVHcloud models support function calling (tool use), which enables the model to interact with external tools and APIs. Models with function calling support include:
- Llama 3.1 8B, Llama 3.3 70B
- All Qwen 3 models and Qwen 2.5 Coder
- GPT-OSS 20B and 120B
- DeepSeek R1 Distill Llama 70B
- Mistral Small 3.2 24B, Mistral Nemo
Function calling is automatically enabled for supported models when using Continue.
```yaml title="config.yaml" name: My Config version: 0.0.1 schema: v1models:
- name: Qwen2.5-Coder-32B-Instruct
provider: ovhcloud
model: qwen2.5-coder-32b
apiKey: <YOUR_AIENDPOINTS_API_KEY>
```
</Tab>
<Tab title="JSON">
```json title="config.json"
{
"models": [
{
"title": "Qwen2.5-Coder-32B-Instruct",
"provider": "ovhcloud",
"model": "qwen2.5-coder-32b",
"apiKey": "<YOUR_AIENDPOINTS_API_KEY>"
}
]
}
```
</Tab>
Here's an example configuration for a model that supports function calling:
```yaml title="config.yaml" name: My Config version: 0.0.1 schema: v1models:
- name: GPT-OSS-120B
provider: ovhcloud
model: gpt-oss-120b
apiKey: <YOUR_AIENDPOINTS_API_KEY>
capabilities:
- tool_use
```
</Tab>
<Tab title="JSON">
```json title="config.json"
{
"models": [
{
"title": "GPT-OSS-120B",
"provider": "ovhcloud",
"model": "gpt-oss-120b",
"apiKey": "<YOUR_AIENDPOINTS_API_KEY>",
"capabilities": ["tool_use"]
}
]
}
```
</Tab>
We recommend configuring bge-multilingual-gemma2 as your embeddings model.
```yaml title="config.yaml" name: My Config version: 0.0.1 schema: v1models: - name: BGE Multilingual Gemma2 provider: ovhcloud model: bge-multilingual-gemma2 apiKey: <YOUR_AIENDPOINTS_API_KEY> roles: - embed
</Tab>
<Tab title="JSON">
```json title="config.json"
{
"embeddingsProvider": {
"provider": "ovhcloud",
"model": "bge-multilingual-gemma2",
"apiKey": "<YOUR_AIENDPOINTS_API_KEY>"
}
}