apps/web/content/docs/faq/9.local-llm-setup.mdx
This guide covers setting up local LLMs for AI features like summaries and notes. If you're looking for local speech-to-text models, see Local Models.
LM Studio is a desktop application that lets you run large language models locally on your computer.
http://127.0.0.1:1234/v1 should work if you haven't changed the portSome local models (e.g., Gemma) do not support tool calling, which Char uses for structured output during summarization. Char automatically detects this and falls back to plain text generation with JSON extraction, so summarization still works with these models. No configuration is needed.
If Char cannot connect to LM Studio:
If you see an error like "Cannot truncate prompt with n_keep >= n_ctx" or "request exceeds the available context size", this means the model's context length is too small for your conversation or transcript.
To fix this:
Ollama is a command-line tool for running large language models locally.
http://127.0.0.1:11434/v1 should workIf Char cannot connect to Ollama:
ollama serve)ollama list)When using local LLMs, your data never leaves your device. For more details on how Char handles data with different AI providers, see AI Models & Data Privacy.