docs/mintlify/docs-mintlify-mig-tmp/ollama.mdx
Ollama lets you run AI models locally on your machine. screenpipe integrates natively with Ollama — no API keys, no cloud, completely private.
# install from https://ollama.com then:
ollama run ministral-3
this downloads the model and starts Ollama. you can use any model — ministral-3 is a good starting point (fast, works on most machines).
that's it. screenpipe talks to Ollama on localhost:11434 automatically.
| model | size | best for |
|---|---|---|
ministral-3 | ~2 GB | fast, general use, recommended starting point |
gemma3:4b | ~3 GB | strong quality for size, good for summaries |
qwen3:4b | ~3 GB | multilingual, good reasoning |
deepseek-r1:8b | ~5 GB | strong reasoning, needs 16 GB+ RAM |
pull any model with:
ollama pull <model-name>
if you're running a custom LLM server (Qwen, vLLM, Text Generation WebUI, etc.), screenpipe auto-detects the endpoint format:
GET {endpoint}/v1/modelsGET {endpoint}/api/tagsif your endpoint uses neither format, you may need to:
/models, /v1/list, etc.)curl {your-endpoint}/path-to-modelsexample: a Qwen server on http://localhost:5000 with OpenAI-compatible API should work automatically. if screenpipe can't find models, verify the server responds to: curl http://localhost:5000/v1/models
"ollama not detected"
ollama servecurl http://localhost:11434/api/tagsmodel not showing in dropdown?
ollama pull ministral-3slow responses?
ministral-3)need help? join our discord — get recommendations on models and configs from the community.