Run Cline with local inference on your machine.Documentation Index
Fetch the complete documentation index at: https://docs.cline.bot/llms.txt
Use this file to discover all available pages before exploring further.
Quick Start
- Install a local runtime (Ollama or LM Studio)
- Start the local server
- In Cline Settings, select the matching provider
- Select a local model
- Enable Use Compact Prompt in Cline Settings → Features
Hardware Requirements
| RAM | Typical local setup |
|---|---|
| 16-32GB | Small/quantized models |
| 32-64GB | Mid-size coding models |
| 64GB+ | Larger models and bigger context windows |
Runtime Options
- Ollama
- LM Studio
1) Install
- Download from ollama.com
- Install for your OS
2) Find popular local models
- Browse the Ollama model catalog: ollama.com/search
- Sort/filter by popularity, model size, and latest updates
- Open any model page and copy the
ollama pullcommand
3) Pull and run a model
4) Configure Cline
- Open Cline Settings
- Select provider: Ollama
- Base URL:
http://localhost:11434 - Select your model from the dropdown
5) Troubleshooting
- Make sure Ollama is running before sending prompts
- If connection fails, verify
http://localhost:11434 - If model is missing, run
ollama pull <model-name>
Recommended Cline Settings for Local Inference
- Enable Use Compact Prompt
- Keep tasks focused (smaller context = faster responses)
- Start a new task when context gets too large

