docs/llm-providers/local.mdx
Running Strix with local models allows for completely offline, privacy-first security assessments. Data never leaves your machine, making this ideal for sensitive internal networks or air-gapped environments.
| Feature | Local Models | Cloud Models (GPT-5/Claude 4.5) |
|---|---|---|
| Privacy | 🔒 Data stays local | Data sent to provider |
| Cost | Free (hardware only) | Pay-per-token |
| Reasoning | Lower (struggles with agents) | State-of-the-art |
| Setup | Complex (GPU required) | Instant |
For critical assessments, we strongly recommend using state-of-the-art cloud models like Claude 4.5 Sonnet or GPT-5. Use local models only when privacy is the absolute priority. </Warning>
Ollama is the easiest way to run local models on macOS, Linux, and Windows.
ollama pull qwen3-vl
export STRIX_LLM="ollama/qwen3-vl"
export LLM_API_BASE="http://localhost:11434"
We recommend these models for the best balance of reasoning and tool use:
Recommended models:
ollama pull qwen3-vl)ollama pull deepseek-v3.1)ollama pull devstral-2)If you use LM Studio, vLLM, or other runners:
export STRIX_LLM="openai/local-model"
export LLM_API_BASE="http://localhost:1234/v1" # Adjust port as needed