docs/customize/model-providers/more/ipex_llm.mdx
IPEX-LLM supports accelerated Ollama backend to be hosted on Intel GPU. Refer to this guide from IPEX-LLM official documentation about how to install and run Ollama serve accelerated by IPEX-LLM on Intel GPU. You can then configure Continue to use the IPEX-LLM accelerated "ollama" provider as follows:
models: - name: IPEX-LLM provider: ollama model: AUTODETECT
</Tab>
<Tab title="JSON">
```json title="config.json"
{
"models": [
{
"title": "IPEX-LLM",
"provider": "ollama",
"model": "AUTODETECT"
}
]
}
If you would like to reach the Ollama service from another machine, make sure you set or export the environment variable OLLAMA_HOST=0.0.0.0 before executing the command ollama serve. Then, in the Continue configuration, set 'apiBase' to correspond with the IP address / port of the remote machine. That is, Continue can be configured to be:
models: - name: IPEX-LLM provider: ollama model: AUTODETECT apiBase: http://your-ollama-service-ip:11434
</Tab>
<Tab title="JSON">
```json title="config.json"
{
"models": [
{
"title": "IPEX-LLM",
"provider": "ollama",
"model": "AUTODETECT",
"apiBase": "http://your-ollama-service-ip:11434"
}
]
}