docs/running-models-locally/overview.mdx
Run Cline with local inference on your machine.
| RAM | Typical local setup |
|---|---|
| 16-32GB | Small/quantized models |
| 32-64GB | Mid-size coding models |
| 64GB+ | Larger models and bigger context windows |
### 2) Find popular local models
- Browse the Ollama model catalog: [ollama.com/search](https://ollama.com/search)
- Sort/filter by popularity, model size, and latest updates
- Open any model page and copy the `ollama pull` command
### 3) Pull and run a model
```bash
ollama pull <model-name>
ollama run <model-name>
```
### 4) Configure Cline
1. Open Cline Settings
2. Select provider: **Ollama**
3. Base URL: `http://localhost:11434`
4. Select your model from the dropdown
### 5) Troubleshooting
- Make sure Ollama is running before sending prompts
- If connection fails, verify `http://localhost:11434`
- If model is missing, run `ollama pull <model-name>`
### 2) Find local models
- Browse the LM Studio model catalog: [lmstudio.ai/models](https://lmstudio.ai/models)
- Filter by model family, size, and capabilities
- Pick a model that matches your hardware
### 3) Download a model
- Open **Discover** and download a model
### 4) Start server
- Open **Developer** tab
- Start server (default: `http://localhost:1234`)
### 5) Configure Cline
1. Open Cline Settings
2. Select provider: **LM Studio**
3. Keep Base URL as `http://localhost:1234`
4. Select your model from the dropdown
### 6) Troubleshooting
- Ensure LM Studio server is running
- Ensure a model is loaded
- If connection fails, verify `http://localhost:1234`