Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.cline.bot/llms.txt

Use this file to discover all available pages before exploring further.

Run Cline with local inference on your machine.

Quick Start

  1. Install a local runtime (Ollama or LM Studio)
  2. Start the local server
  3. In Cline Settings, select the matching provider
  4. Select a local model
  5. Enable Use Compact Prompt in Cline Settings → Features

Hardware Requirements

RAMTypical local setup
16-32GBSmall/quantized models
32-64GBMid-size coding models
64GB+Larger models and bigger context windows

Runtime Options

1) Install

  • Browse the Ollama model catalog: ollama.com/search
  • Sort/filter by popularity, model size, and latest updates
  • Open any model page and copy the ollama pull command

3) Pull and run a model

ollama pull <model-name>
ollama run <model-name>

4) Configure Cline

  1. Open Cline Settings
  2. Select provider: Ollama
  3. Base URL: http://localhost:11434
  4. Select your model from the dropdown

5) Troubleshooting

  • Make sure Ollama is running before sending prompts
  • If connection fails, verify http://localhost:11434
  • If model is missing, run ollama pull <model-name>
  • Enable Use Compact Prompt
  • Keep tasks focused (smaller context = faster responses)
  • Start a new task when context gets too large