Back to Cline

Local models

docs/running-models-locally/overview.mdx

3.83.02.4 KB
Original Source

Run Cline with local inference on your machine.

Quick Start

  1. Install a local runtime (Ollama or LM Studio)
  2. Start the local server
  3. In Cline Settings, select the matching provider
  4. Select a local model
  5. Enable Use Compact Prompt in Cline Settings → Features

Hardware Requirements

RAMTypical local setup
16-32GBSmall/quantized models
32-64GBMid-size coding models
64GB+Larger models and bigger context windows

Runtime Options

<Tabs> <Tab title="Ollama"> ### 1) Install - Download from [ollama.com](https://ollama.com) - Install for your OS
### 2) Find popular local models
- Browse the Ollama model catalog: [ollama.com/search](https://ollama.com/search)
- Sort/filter by popularity, model size, and latest updates
- Open any model page and copy the `ollama pull` command

### 3) Pull and run a model
```bash
ollama pull <model-name>
ollama run <model-name>
```

### 4) Configure Cline
1. Open Cline Settings
2. Select provider: **Ollama**
3. Base URL: `http://localhost:11434`
4. Select your model from the dropdown

### 5) Troubleshooting
- Make sure Ollama is running before sending prompts
- If connection fails, verify `http://localhost:11434`
- If model is missing, run `ollama pull <model-name>`
</Tab> <Tab title="LM Studio"> ### 1) Install - Download from [lmstudio.ai](https://lmstudio.ai) - Install and launch the app
### 2) Find local models
- Browse the LM Studio model catalog: [lmstudio.ai/models](https://lmstudio.ai/models)
- Filter by model family, size, and capabilities
- Pick a model that matches your hardware

### 3) Download a model
- Open **Discover** and download a model

### 4) Start server
- Open **Developer** tab
- Start server (default: `http://localhost:1234`)

### 5) Configure Cline
1. Open Cline Settings
2. Select provider: **LM Studio**
3. Keep Base URL as `http://localhost:1234`
4. Select your model from the dropdown

### 6) Troubleshooting
- Ensure LM Studio server is running
- Ensure a model is loaded
- If connection fails, verify `http://localhost:1234`
</Tab> </Tabs>
  • Enable Use Compact Prompt
  • Keep tasks focused (smaller context = faster responses)
  • Start a new task when context gets too large