Skip to main content
💻 Workflow Template

Local Coding Assistant

Set up a private AI coding assistant using Ollama and a code-optimised model — your Copilot alternative with zero cloud dependency.

Setup time

20–40 minutes

Min hardware

8 GB VRAM (RTX 3060 / RX 6600 XT)

Software

Ollama + Continue (VS Code extension)

Recommended model

Qwen2.5-Coder 7B

1

Install Ollama

Download Ollama from ollama.com and install it on your system.

2

Pull a code model

Run: ollama pull qwen2.5-coder:7b. Qwen2.5-Coder outperforms Codellama on most benchmarks.

Tip:

For lower VRAM (6 GB): ollama pull qwen2.5-coder:3b

3

Install Continue in VS Code

Open VS Code, go to Extensions and search for Continue. Install the Continue extension (continueDev.continue). It connects to your local Ollama instance automatically.

4

Configure Continue

Open the Continue sidebar, click the settings icon, and add your model: select Ollama as provider and choose qwen2.5-coder:7b.

5

Test code completion

Open any code file, highlight a function, and press Ctrl+I (or Cmd+I on Mac) to invoke the AI. Ask it to refactor, explain, or complete code.

Affiliate disclosure: links to Amazon may earn us a commission at no extra cost to you.

Compatible GPUs for this workflow

For smooth local AI inference with this workflow, these NVIDIA RTX GPUs deliver the best experience.

Prices and availability may change.